var/home/core/zuul-output/0000755000175000017500000000000015111011414014512 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111017413015463 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004211306015111017401017661 0ustar rootrootNov 24 08:15:24 crc systemd[1]: Starting Kubernetes Kubelet... Nov 24 08:15:25 crc restorecon[4636]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 24 08:15:25 crc restorecon[4636]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 24 08:15:26 crc kubenswrapper[4831]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.620960 4831 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.629939 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.629976 4831 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.629984 4831 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.629993 4831 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630000 4831 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630007 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630013 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630019 4831 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630026 4831 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630032 4831 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630038 4831 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630044 4831 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630049 4831 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630055 4831 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630060 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630065 4831 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630071 4831 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630076 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630081 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630087 4831 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630094 4831 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630100 4831 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630106 4831 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630112 4831 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630118 4831 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630126 4831 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630133 4831 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630141 4831 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630178 4831 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630186 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630192 4831 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630198 4831 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630203 4831 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630209 4831 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630216 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630222 4831 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630227 4831 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630233 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630239 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630244 4831 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630249 4831 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630255 4831 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630260 4831 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630266 4831 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630271 4831 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630279 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630284 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630290 4831 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630295 4831 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630300 4831 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630306 4831 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630311 4831 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630334 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630340 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630345 4831 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630350 4831 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630356 4831 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630362 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630368 4831 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630373 4831 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630379 4831 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630387 4831 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630396 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630402 4831 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630407 4831 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630412 4831 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630417 4831 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630425 4831 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630431 4831 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630437 4831 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.630443 4831 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631248 4831 flags.go:64] FLAG: --address="0.0.0.0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631268 4831 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631281 4831 flags.go:64] FLAG: --anonymous-auth="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631289 4831 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631297 4831 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631303 4831 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631312 4831 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631348 4831 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631354 4831 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631360 4831 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631367 4831 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631375 4831 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631381 4831 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631387 4831 flags.go:64] FLAG: --cgroup-root="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631393 4831 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631398 4831 flags.go:64] FLAG: --client-ca-file="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631405 4831 flags.go:64] FLAG: --cloud-config="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631410 4831 flags.go:64] FLAG: --cloud-provider="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631417 4831 flags.go:64] FLAG: --cluster-dns="[]" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631424 4831 flags.go:64] FLAG: --cluster-domain="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631430 4831 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631437 4831 flags.go:64] FLAG: --config-dir="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631443 4831 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631450 4831 flags.go:64] FLAG: --container-log-max-files="5" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631458 4831 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631464 4831 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631471 4831 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631478 4831 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631483 4831 flags.go:64] FLAG: --contention-profiling="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631489 4831 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631495 4831 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631502 4831 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631509 4831 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631516 4831 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631522 4831 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631529 4831 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631535 4831 flags.go:64] FLAG: --enable-load-reader="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631542 4831 flags.go:64] FLAG: --enable-server="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631548 4831 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631557 4831 flags.go:64] FLAG: --event-burst="100" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631564 4831 flags.go:64] FLAG: --event-qps="50" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631570 4831 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631577 4831 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631583 4831 flags.go:64] FLAG: --eviction-hard="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631591 4831 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631597 4831 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631602 4831 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631610 4831 flags.go:64] FLAG: --eviction-soft="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631616 4831 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631622 4831 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631628 4831 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631634 4831 flags.go:64] FLAG: --experimental-mounter-path="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631640 4831 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631646 4831 flags.go:64] FLAG: --fail-swap-on="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631652 4831 flags.go:64] FLAG: --feature-gates="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631659 4831 flags.go:64] FLAG: --file-check-frequency="20s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631666 4831 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631675 4831 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631682 4831 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631687 4831 flags.go:64] FLAG: --healthz-port="10248" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631693 4831 flags.go:64] FLAG: --help="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631699 4831 flags.go:64] FLAG: --hostname-override="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631705 4831 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631711 4831 flags.go:64] FLAG: --http-check-frequency="20s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631717 4831 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631723 4831 flags.go:64] FLAG: --image-credential-provider-config="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631729 4831 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631734 4831 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631740 4831 flags.go:64] FLAG: --image-service-endpoint="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631745 4831 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631751 4831 flags.go:64] FLAG: --kube-api-burst="100" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631757 4831 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631763 4831 flags.go:64] FLAG: --kube-api-qps="50" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631769 4831 flags.go:64] FLAG: --kube-reserved="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631775 4831 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631781 4831 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631787 4831 flags.go:64] FLAG: --kubelet-cgroups="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631793 4831 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631803 4831 flags.go:64] FLAG: --lock-file="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631809 4831 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631816 4831 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631822 4831 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631839 4831 flags.go:64] FLAG: --log-json-split-stream="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631847 4831 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631853 4831 flags.go:64] FLAG: --log-text-split-stream="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631859 4831 flags.go:64] FLAG: --logging-format="text" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631865 4831 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631872 4831 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631878 4831 flags.go:64] FLAG: --manifest-url="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631885 4831 flags.go:64] FLAG: --manifest-url-header="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631895 4831 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631902 4831 flags.go:64] FLAG: --max-open-files="1000000" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631910 4831 flags.go:64] FLAG: --max-pods="110" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631917 4831 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631923 4831 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631930 4831 flags.go:64] FLAG: --memory-manager-policy="None" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631936 4831 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631942 4831 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631949 4831 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631955 4831 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631971 4831 flags.go:64] FLAG: --node-status-max-images="50" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631977 4831 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.631995 4831 flags.go:64] FLAG: --oom-score-adj="-999" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632002 4831 flags.go:64] FLAG: --pod-cidr="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632008 4831 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632020 4831 flags.go:64] FLAG: --pod-manifest-path="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632027 4831 flags.go:64] FLAG: --pod-max-pids="-1" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632034 4831 flags.go:64] FLAG: --pods-per-core="0" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632041 4831 flags.go:64] FLAG: --port="10250" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632048 4831 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632056 4831 flags.go:64] FLAG: --provider-id="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632062 4831 flags.go:64] FLAG: --qos-reserved="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632069 4831 flags.go:64] FLAG: --read-only-port="10255" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632075 4831 flags.go:64] FLAG: --register-node="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632081 4831 flags.go:64] FLAG: --register-schedulable="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632087 4831 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632098 4831 flags.go:64] FLAG: --registry-burst="10" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632104 4831 flags.go:64] FLAG: --registry-qps="5" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632110 4831 flags.go:64] FLAG: --reserved-cpus="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632117 4831 flags.go:64] FLAG: --reserved-memory="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632125 4831 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632131 4831 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632137 4831 flags.go:64] FLAG: --rotate-certificates="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632143 4831 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632149 4831 flags.go:64] FLAG: --runonce="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632155 4831 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632162 4831 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632168 4831 flags.go:64] FLAG: --seccomp-default="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632174 4831 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632181 4831 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632187 4831 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632194 4831 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632200 4831 flags.go:64] FLAG: --storage-driver-password="root" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632206 4831 flags.go:64] FLAG: --storage-driver-secure="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632212 4831 flags.go:64] FLAG: --storage-driver-table="stats" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632218 4831 flags.go:64] FLAG: --storage-driver-user="root" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632225 4831 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632231 4831 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632237 4831 flags.go:64] FLAG: --system-cgroups="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632244 4831 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632254 4831 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632261 4831 flags.go:64] FLAG: --tls-cert-file="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632268 4831 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632275 4831 flags.go:64] FLAG: --tls-min-version="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632281 4831 flags.go:64] FLAG: --tls-private-key-file="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632287 4831 flags.go:64] FLAG: --topology-manager-policy="none" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632293 4831 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632300 4831 flags.go:64] FLAG: --topology-manager-scope="container" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632307 4831 flags.go:64] FLAG: --v="2" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632362 4831 flags.go:64] FLAG: --version="false" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632372 4831 flags.go:64] FLAG: --vmodule="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632380 4831 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.632387 4831 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632691 4831 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632701 4831 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632709 4831 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632715 4831 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632721 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632726 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632732 4831 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632737 4831 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632743 4831 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632748 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632753 4831 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632758 4831 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632764 4831 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632769 4831 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632773 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632778 4831 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632783 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632788 4831 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632793 4831 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632798 4831 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632804 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632810 4831 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632815 4831 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632820 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632826 4831 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632833 4831 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632838 4831 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632844 4831 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632849 4831 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632854 4831 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632859 4831 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632864 4831 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632869 4831 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632877 4831 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632882 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632888 4831 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632895 4831 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632902 4831 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632910 4831 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632917 4831 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632924 4831 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632931 4831 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632937 4831 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632943 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632948 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632954 4831 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632961 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632967 4831 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632973 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632978 4831 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632984 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632990 4831 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.632996 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633002 4831 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633007 4831 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633012 4831 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633018 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633024 4831 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633029 4831 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633034 4831 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633039 4831 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633045 4831 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633050 4831 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633055 4831 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633061 4831 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633066 4831 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633072 4831 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633077 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633082 4831 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633087 4831 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.633094 4831 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.633113 4831 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.645252 4831 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.645339 4831 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645422 4831 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645431 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645436 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645443 4831 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645451 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645456 4831 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645460 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645465 4831 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645470 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645475 4831 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645479 4831 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645485 4831 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645489 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645494 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645498 4831 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645503 4831 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645508 4831 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645512 4831 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645516 4831 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645521 4831 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645525 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645529 4831 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645533 4831 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645538 4831 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645542 4831 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645547 4831 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645551 4831 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645555 4831 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645559 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645564 4831 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645567 4831 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645572 4831 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645576 4831 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645580 4831 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645586 4831 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645590 4831 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645594 4831 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645600 4831 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645606 4831 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645610 4831 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645615 4831 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645620 4831 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645625 4831 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645629 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645634 4831 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645639 4831 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645643 4831 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645650 4831 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645656 4831 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645660 4831 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645665 4831 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645669 4831 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645675 4831 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645680 4831 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645684 4831 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645688 4831 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645694 4831 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645700 4831 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645704 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645708 4831 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645713 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645717 4831 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645722 4831 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645726 4831 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645731 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645735 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645739 4831 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645743 4831 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645747 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645752 4831 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645757 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.645765 4831 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645907 4831 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645914 4831 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645920 4831 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645926 4831 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645931 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645936 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645941 4831 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645947 4831 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645952 4831 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645957 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645961 4831 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645966 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645972 4831 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645977 4831 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645981 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645986 4831 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645990 4831 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.645995 4831 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646001 4831 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646006 4831 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646012 4831 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646018 4831 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646022 4831 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646027 4831 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646032 4831 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646037 4831 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646042 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646047 4831 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646052 4831 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646056 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646060 4831 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646065 4831 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646070 4831 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646074 4831 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646079 4831 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646084 4831 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646088 4831 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646092 4831 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646096 4831 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646100 4831 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646104 4831 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646108 4831 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646114 4831 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646118 4831 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646123 4831 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646128 4831 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646133 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646138 4831 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646142 4831 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646146 4831 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646150 4831 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646154 4831 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646158 4831 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646163 4831 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646167 4831 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646171 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646175 4831 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646179 4831 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646184 4831 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646189 4831 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646193 4831 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646197 4831 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646201 4831 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646205 4831 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646211 4831 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646216 4831 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646220 4831 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646224 4831 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646228 4831 feature_gate.go:330] unrecognized feature gate: Example Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646232 4831 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.646237 4831 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.646245 4831 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.646515 4831 server.go:940] "Client rotation is on, will bootstrap in background" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.651261 4831 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.651418 4831 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.654147 4831 server.go:997] "Starting client certificate rotation" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.654173 4831 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.654346 4831 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-10 09:25:23.549863982 +0000 UTC Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.654430 4831 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 385h9m56.895437288s for next certificate rotation Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.689631 4831 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.691347 4831 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.712015 4831 log.go:25] "Validated CRI v1 runtime API" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.745677 4831 log.go:25] "Validated CRI v1 image API" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.748554 4831 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.757934 4831 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-24-08-09-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.757964 4831 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.779790 4831 manager.go:217] Machine: {Timestamp:2025-11-24 08:15:26.776254384 +0000 UTC m=+0.651399617 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f42275a4-fa6b-4c08-99eb-479953ecb2b1 BootID:59b72184-9079-421a-8fa8-bad568411b37 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8f:90:71 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8f:90:71 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9d:da:1b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:57:20:9b Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:92:2f:47 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:6a:ca:89 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:58:c0:07 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:72:87:7e:63:9f:02 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:36:b0:f0:72:5e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.780111 4831 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.780291 4831 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.782129 4831 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.782391 4831 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.782434 4831 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.784278 4831 topology_manager.go:138] "Creating topology manager with none policy" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.784301 4831 container_manager_linux.go:303] "Creating device plugin manager" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.784813 4831 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.784842 4831 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.785034 4831 state_mem.go:36] "Initialized new in-memory state store" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.785115 4831 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.788851 4831 kubelet.go:418] "Attempting to sync node with API server" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.788877 4831 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.788900 4831 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.788914 4831 kubelet.go:324] "Adding apiserver pod source" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.788928 4831 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.793976 4831 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.795028 4831 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.796285 4831 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.796979 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.797042 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.797028 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.797146 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798776 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798800 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798808 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798816 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798828 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798836 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798844 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798861 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798872 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798879 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798889 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.798897 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.802151 4831 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.802544 4831 server.go:1280] "Started kubelet" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.804628 4831 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.804724 4831 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:26 crc systemd[1]: Started Kubernetes Kubelet. Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.804626 4831 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.805397 4831 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.806952 4831 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807006 4831 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807201 4831 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 14:24:54.154105749 +0000 UTC Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807242 4831 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 30h9m27.346866445s for next certificate rotation Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.807372 4831 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807393 4831 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807673 4831 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.807415 4831 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.808166 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.809016 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.809838 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="200ms" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810389 4831 factory.go:153] Registering CRI-O factory Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810413 4831 factory.go:221] Registration of the crio container factory successfully Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810480 4831 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810489 4831 factory.go:55] Registering systemd factory Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810495 4831 factory.go:221] Registration of the systemd container factory successfully Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810513 4831 factory.go:103] Registering Raw factory Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.810525 4831 manager.go:1196] Started watching for new ooms in manager Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.811076 4831 manager.go:319] Starting recovery of all containers Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.811635 4831 server.go:460] "Adding debug handlers to kubelet server" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.835595 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.835670 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.835687 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839179 4831 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839236 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839261 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839280 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839311 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839344 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839363 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839403 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839417 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839435 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839450 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839490 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839505 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839526 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839562 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839634 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839653 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839669 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839688 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839732 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839747 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839764 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839808 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839825 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839844 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839880 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839898 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839914 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.839931 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840003 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840021 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840057 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840074 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840090 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840107 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840148 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840165 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840218 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840240 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840283 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840299 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840331 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840346 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840362 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840377 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840395 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840411 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840452 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840472 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840487 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840508 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840524 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840539 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840579 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840595 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840610 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840624 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840636 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840652 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840667 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840681 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840697 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840713 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840729 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840745 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840762 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840777 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840792 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840835 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840853 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840869 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840886 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840902 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840915 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840956 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840971 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.840984 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841000 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841016 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841030 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841046 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841060 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841079 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841095 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841109 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841129 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841146 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841162 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841211 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841228 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841243 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841258 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841275 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841294 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841308 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841347 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841360 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841380 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841393 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841407 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841423 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841438 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841457 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841471 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841505 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841548 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841572 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841589 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841611 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841630 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841644 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841658 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841673 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841685 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841698 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841711 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841724 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841737 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841750 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841765 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841779 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841790 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841802 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841815 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841828 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841841 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841854 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841869 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841881 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841893 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841908 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841921 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841934 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841948 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841960 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841974 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.841991 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842004 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842022 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842036 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842049 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842062 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842074 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842086 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842099 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842111 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842123 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842139 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842152 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842164 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842179 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842195 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842209 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842221 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842234 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842248 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842265 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842281 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842296 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842307 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842338 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842355 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842368 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842382 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842394 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842409 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842420 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842432 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842444 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842455 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842469 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842482 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842496 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842513 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842525 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842535 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842548 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842561 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842575 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842587 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842600 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842612 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842626 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842644 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842659 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842676 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842690 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842705 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842719 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842732 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842745 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842759 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842773 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842784 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842797 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842810 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842824 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842837 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842849 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842886 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842900 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842912 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842925 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842941 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842953 4831 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842972 4831 reconstruct.go:97] "Volume reconstruction finished" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.842980 4831 reconciler.go:26] "Reconciler: start to sync state" Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.844719 4831 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.136:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ae34dabc0fee0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-24 08:15:26.802509536 +0000 UTC m=+0.677654689,LastTimestamp:2025-11-24 08:15:26.802509536 +0000 UTC m=+0.677654689,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.853256 4831 manager.go:324] Recovery completed Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.871498 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.873462 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.873564 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.873589 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.874975 4831 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.874995 4831 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.875016 4831 state_mem.go:36] "Initialized new in-memory state store" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.890042 4831 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.892063 4831 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.892114 4831 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.892147 4831 kubelet.go:2335] "Starting kubelet main sync loop" Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.892204 4831 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 24 08:15:26 crc kubenswrapper[4831]: W1124 08:15:26.893619 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.893698 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.905450 4831 policy_none.go:49] "None policy: Start" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.906285 4831 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.906338 4831 state_mem.go:35] "Initializing new in-memory state store" Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.907478 4831 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.978546 4831 manager.go:334] "Starting Device Plugin manager" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.978598 4831 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.978612 4831 server.go:79] "Starting device plugin registration server" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.979201 4831 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.979226 4831 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.979456 4831 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.979536 4831 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.979546 4831 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 24 08:15:26 crc kubenswrapper[4831]: E1124 08:15:26.986267 4831 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.992434 4831 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.992608 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.993821 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.993862 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.993875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994019 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994274 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994308 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994889 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994892 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994916 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994948 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.994961 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995035 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995207 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995375 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995581 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995609 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995619 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995720 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995836 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.995871 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996427 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996447 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996467 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996524 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996533 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996593 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996689 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.996709 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997331 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997352 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997360 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997370 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997386 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997394 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997512 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.997549 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998061 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998075 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998185 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:26 crc kubenswrapper[4831]: I1124 08:15:26.998235 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.010820 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="400ms" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045242 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045285 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045340 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045367 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045407 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045446 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045472 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045495 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045516 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045536 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045557 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045578 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045604 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045626 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.045650 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.079780 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.080735 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.080774 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.080786 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.080812 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.081259 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146756 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146847 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146869 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146909 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146915 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146930 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147011 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147022 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147022 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147078 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147020 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.146930 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147077 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147113 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147177 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147134 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147201 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147223 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147250 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147272 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147347 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147293 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147384 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147390 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147395 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147425 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147500 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.147530 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.281871 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.283392 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.283429 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.283437 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.283458 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.284000 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.339138 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.348228 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.368880 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.382227 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.387662 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.390262 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-272ed2ffc271d4de0ddc5338670c5fc5e9b79d00ce293716a333ec9475ecc0c7 WatchSource:0}: Error finding container 272ed2ffc271d4de0ddc5338670c5fc5e9b79d00ce293716a333ec9475ecc0c7: Status 404 returned error can't find the container with id 272ed2ffc271d4de0ddc5338670c5fc5e9b79d00ce293716a333ec9475ecc0c7 Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.393796 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4c9b0fac9baa19b58cbf9f37a56c14db8c96da9a6765869d8a86291dee7c8548 WatchSource:0}: Error finding container 4c9b0fac9baa19b58cbf9f37a56c14db8c96da9a6765869d8a86291dee7c8548: Status 404 returned error can't find the container with id 4c9b0fac9baa19b58cbf9f37a56c14db8c96da9a6765869d8a86291dee7c8548 Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.401779 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-328ab142d1ed2ad17de29acce5bfc311ab1407e00eb427b2bb7a3e195e746d18 WatchSource:0}: Error finding container 328ab142d1ed2ad17de29acce5bfc311ab1407e00eb427b2bb7a3e195e746d18: Status 404 returned error can't find the container with id 328ab142d1ed2ad17de29acce5bfc311ab1407e00eb427b2bb7a3e195e746d18 Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.408297 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-4e14b54feab202ec5de658d3cece3877d0bf80408ee91dd1f1a4e84805df66b5 WatchSource:0}: Error finding container 4e14b54feab202ec5de658d3cece3877d0bf80408ee91dd1f1a4e84805df66b5: Status 404 returned error can't find the container with id 4e14b54feab202ec5de658d3cece3877d0bf80408ee91dd1f1a4e84805df66b5 Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.411286 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="800ms" Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.661592 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.661690 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.685022 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.686654 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.686693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.686705 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.686731 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.687174 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.783010 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.783115 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.806497 4831 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:27 crc kubenswrapper[4831]: W1124 08:15:27.862829 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:27 crc kubenswrapper[4831]: E1124 08:15:27.862920 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.898446 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"4e14b54feab202ec5de658d3cece3877d0bf80408ee91dd1f1a4e84805df66b5"} Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.899723 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"328ab142d1ed2ad17de29acce5bfc311ab1407e00eb427b2bb7a3e195e746d18"} Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.901404 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4c9b0fac9baa19b58cbf9f37a56c14db8c96da9a6765869d8a86291dee7c8548"} Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.902381 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"272ed2ffc271d4de0ddc5338670c5fc5e9b79d00ce293716a333ec9475ecc0c7"} Nov 24 08:15:27 crc kubenswrapper[4831]: I1124 08:15:27.903399 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4ac47a60492ee1b10c45e26e05afaac760239417d074ebe7e23ca5b3f688aca5"} Nov 24 08:15:28 crc kubenswrapper[4831]: W1124 08:15:28.173757 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:28 crc kubenswrapper[4831]: E1124 08:15:28.173875 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:28 crc kubenswrapper[4831]: E1124 08:15:28.212163 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="1.6s" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.487748 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.489980 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.490038 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.490051 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.490082 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:28 crc kubenswrapper[4831]: E1124 08:15:28.490849 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.806048 4831 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.906793 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.906854 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.906870 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.906882 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.906814 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.907921 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.907941 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.907950 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.910177 4831 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="75c729293fa4983e9b3bcc754cf547838f43edaccd4965b4873ef0b8704e222e" exitCode=0 Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.910214 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"75c729293fa4983e9b3bcc754cf547838f43edaccd4965b4873ef0b8704e222e"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.910266 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.911229 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.911298 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.911313 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.912399 4831 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63" exitCode=0 Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.912486 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.912761 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.913619 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.913660 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.913676 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.915555 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.916886 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.916948 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.916963 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.917079 4831 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89" exitCode=0 Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.917182 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.917225 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.918170 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.918211 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.918227 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.922254 4831 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50" exitCode=0 Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.922355 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50"} Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.922908 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.926087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.926128 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.926142 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:28 crc kubenswrapper[4831]: I1124 08:15:28.988943 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.184581 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.806189 4831 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:29 crc kubenswrapper[4831]: E1124 08:15:29.813785 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.136:6443: connect: connection refused" interval="3.2s" Nov 24 08:15:29 crc kubenswrapper[4831]: W1124 08:15:29.890589 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:29 crc kubenswrapper[4831]: E1124 08:15:29.890695 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.926806 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.926860 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.926872 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.926881 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.926893 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.928260 4831 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e" exitCode=0 Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.928334 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.928446 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.929209 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.929231 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.929239 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.932417 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.932469 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.932481 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.932436 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.933462 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.933484 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.933493 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.934441 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.934522 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.934624 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e053641ccd44b22bf5192db01a85ff4a7d7903d0d25aeab4c99e5eba369d6ca1"} Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.935261 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.935280 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.935288 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.938656 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.938797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:29 crc kubenswrapper[4831]: I1124 08:15:29.938814 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.091422 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.092627 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.092657 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.092666 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.092726 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:30 crc kubenswrapper[4831]: E1124 08:15:30.093182 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.136:6443: connect: connection refused" node="crc" Nov 24 08:15:30 crc kubenswrapper[4831]: W1124 08:15:30.121702 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:30 crc kubenswrapper[4831]: E1124 08:15:30.121805 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:30 crc kubenswrapper[4831]: W1124 08:15:30.239535 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.136:6443: connect: connection refused Nov 24 08:15:30 crc kubenswrapper[4831]: E1124 08:15:30.239638 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.136:6443: connect: connection refused" logger="UnhandledError" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939406 4831 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574" exitCode=0 Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939507 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939521 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939565 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939577 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574"} Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939662 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940111 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.939565 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940282 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940311 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940345 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940479 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940504 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940514 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940523 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940544 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940555 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940911 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940944 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940951 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.940998 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:30 crc kubenswrapper[4831]: I1124 08:15:30.941009 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.654820 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.945440 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.945973 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb"} Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946021 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16"} Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946031 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec"} Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946039 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a"} Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946100 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946146 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946387 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946411 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.946420 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.947545 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.947756 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:31 crc kubenswrapper[4831]: I1124 08:15:31.947768 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.184737 4831 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.185063 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.272024 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.817518 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.818418 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.821024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.821091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.821111 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.827795 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.957969 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473"} Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.958104 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.958157 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.958220 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.958223 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.959691 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.959768 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.959796 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960115 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960137 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960084 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960353 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:32 crc kubenswrapper[4831]: I1124 08:15:32.960390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.294199 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.296118 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.296168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.296182 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.296209 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.498634 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.881852 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.961690 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.961791 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.961899 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.963825 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.963896 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.963926 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.964024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.964074 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:33 crc kubenswrapper[4831]: I1124 08:15:33.964098 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:34 crc kubenswrapper[4831]: I1124 08:15:34.964087 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:34 crc kubenswrapper[4831]: I1124 08:15:34.964945 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:34 crc kubenswrapper[4831]: I1124 08:15:34.964972 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:34 crc kubenswrapper[4831]: I1124 08:15:34.964982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:35 crc kubenswrapper[4831]: I1124 08:15:35.456960 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:35 crc kubenswrapper[4831]: I1124 08:15:35.457167 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:35 crc kubenswrapper[4831]: I1124 08:15:35.458229 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:35 crc kubenswrapper[4831]: I1124 08:15:35.458268 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:35 crc kubenswrapper[4831]: I1124 08:15:35.458278 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:36 crc kubenswrapper[4831]: I1124 08:15:36.621817 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:36 crc kubenswrapper[4831]: I1124 08:15:36.621994 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:36 crc kubenswrapper[4831]: I1124 08:15:36.623119 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:36 crc kubenswrapper[4831]: I1124 08:15:36.623165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:36 crc kubenswrapper[4831]: I1124 08:15:36.623174 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:36 crc kubenswrapper[4831]: E1124 08:15:36.986382 4831 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 24 08:15:39 crc kubenswrapper[4831]: I1124 08:15:39.606153 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 24 08:15:39 crc kubenswrapper[4831]: I1124 08:15:39.606365 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:39 crc kubenswrapper[4831]: I1124 08:15:39.607603 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:39 crc kubenswrapper[4831]: I1124 08:15:39.607639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:39 crc kubenswrapper[4831]: I1124 08:15:39.607648 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:40 crc kubenswrapper[4831]: W1124 08:15:40.705085 4831 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.705450 4831 trace.go:236] Trace[568256889]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:15:30.703) (total time: 10001ms): Nov 24 08:15:40 crc kubenswrapper[4831]: Trace[568256889]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (08:15:40.705) Nov 24 08:15:40 crc kubenswrapper[4831]: Trace[568256889]: [10.001385946s] [10.001385946s] END Nov 24 08:15:40 crc kubenswrapper[4831]: E1124 08:15:40.705563 4831 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.806963 4831 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.840303 4831 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37750->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.840416 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37750->192.168.126.11:17697: read: connection reset by peer" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.983546 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.985360 4831 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351" exitCode=255 Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.985424 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351"} Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.985645 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.986858 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.986894 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.986906 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:40 crc kubenswrapper[4831]: I1124 08:15:40.987466 4831 scope.go:117] "RemoveContainer" containerID="f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.108154 4831 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.108285 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.116623 4831 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.116687 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.989828 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.991363 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117"} Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.991539 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.992503 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.992536 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:41 crc kubenswrapper[4831]: I1124 08:15:41.992551 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:42 crc kubenswrapper[4831]: I1124 08:15:42.185823 4831 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 08:15:42 crc kubenswrapper[4831]: I1124 08:15:42.185882 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.504886 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.505068 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.505187 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.507596 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.507652 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.507670 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.511191 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.996041 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.996910 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.996968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:43 crc kubenswrapper[4831]: I1124 08:15:43.996984 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:44 crc kubenswrapper[4831]: I1124 08:15:44.998215 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:44 crc kubenswrapper[4831]: I1124 08:15:44.999299 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:44 crc kubenswrapper[4831]: I1124 08:15:44.999369 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:44 crc kubenswrapper[4831]: I1124 08:15:44.999383 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.461478 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.461629 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.462629 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.462662 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.462672 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:45 crc kubenswrapper[4831]: I1124 08:15:45.938357 4831 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.114307 4831 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.115385 4831 trace.go:236] Trace[1455866224]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:15:35.571) (total time: 10543ms): Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[1455866224]: ---"Objects listed" error: 10543ms (08:15:46.115) Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[1455866224]: [10.543427002s] [10.543427002s] END Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.115412 4831 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.115798 4831 trace.go:236] Trace[801952851]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:15:34.213) (total time: 11902ms): Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[801952851]: ---"Objects listed" error: 11902ms (08:15:46.115) Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[801952851]: [11.902107438s] [11.902107438s] END Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.115814 4831 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.117305 4831 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.117799 4831 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.118688 4831 trace.go:236] Trace[367140599]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Nov-2025 08:15:34.957) (total time: 11161ms): Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[367140599]: ---"Objects listed" error: 11161ms (08:15:46.118) Nov 24 08:15:46 crc kubenswrapper[4831]: Trace[367140599]: [11.161609944s] [11.161609944s] END Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.118705 4831 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.800394 4831 apiserver.go:52] "Watching apiserver" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.805887 4831 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806215 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-556kb","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806625 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806795 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806874 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806897 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806654 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.806697 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.807152 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.807153 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.807243 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.807289 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.809019 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.810398 4831 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811348 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811436 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811451 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811565 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811620 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.811780 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.813045 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.814592 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.814750 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.814815 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.814838 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821829 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821874 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821905 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821928 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821952 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821975 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.821997 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822019 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822044 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822088 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822128 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822131 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822151 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822210 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822230 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822246 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822263 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822277 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822294 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822309 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822340 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822354 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822368 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822382 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822398 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822415 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822428 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822431 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822520 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822536 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822551 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822565 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822580 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822594 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822632 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822658 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822676 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822692 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822708 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822722 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822741 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822755 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822769 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822786 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822800 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822813 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822828 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822842 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822857 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822859 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822872 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822927 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822951 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822976 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.822998 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823006 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823025 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823049 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823073 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823123 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823141 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823144 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823173 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823189 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823204 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823220 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823234 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823289 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823304 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823341 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823357 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823376 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823392 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823407 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823421 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823437 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823454 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823470 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823486 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823501 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823518 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823534 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823551 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823566 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823581 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823596 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823611 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823626 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823641 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823657 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823672 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823719 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823733 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823747 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823763 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823777 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823792 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823807 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823822 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823836 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823851 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823866 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823882 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823896 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823911 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823925 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823940 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823955 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823970 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824002 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824017 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824031 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824045 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824060 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824075 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824089 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824104 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824118 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824134 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824151 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824166 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824182 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824197 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824534 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824550 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824567 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824581 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824596 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824611 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824626 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824641 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824656 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824671 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824690 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824705 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824722 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824736 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824752 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824767 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824782 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824798 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824814 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824829 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824850 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824865 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824880 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824895 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824910 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824926 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824941 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824958 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824975 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824992 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825008 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825023 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825039 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825055 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825071 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825087 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825102 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825121 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825136 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825151 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825166 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825197 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825212 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825228 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825245 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825261 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825277 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825292 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825308 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825337 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825352 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825368 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825384 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825398 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825414 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825430 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825446 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825462 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825478 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825494 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825512 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825527 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825541 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825556 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825572 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825588 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825605 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825622 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825638 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825656 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825671 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825687 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825702 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825717 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825733 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825749 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825773 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825789 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825805 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830668 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830717 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830751 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht772\" (UniqueName: \"kubernetes.io/projected/f8de0f5f-48af-4be5-a67e-577d6f2897f9-kube-api-access-ht772\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830778 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830802 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830826 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830936 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830961 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.830988 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831008 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f8de0f5f-48af-4be5-a67e-577d6f2897f9-hosts-file\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831212 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831274 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831464 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831630 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831749 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831778 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832523 4831 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832539 4831 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832551 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832573 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832587 4831 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832599 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832610 4831 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.823515 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824297 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824507 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824770 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.824944 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825205 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.825579 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.831489 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832003 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832126 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832311 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832663 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.832823 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.833558 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.833889 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.834044 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.834457 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.834817 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.835271 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838678 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.835517 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.835768 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.835995 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.836713 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838125 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838178 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838433 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838892 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.838961 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839376 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839490 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839514 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839726 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839760 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839930 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.839955 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840062 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840378 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840399 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840699 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840708 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.840773 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.841424 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.841693 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.841773 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.841781 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.842442 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.842607 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.842708 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.845332 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.846436 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.846707 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.847012 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.847290 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.847503 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.847648 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.847839 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.848676 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.848782 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.848826 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.848913 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.848993 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.849284 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.849296 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.849533 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.849660 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.849971 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.850232 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.850832 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.850862 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.851225 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.851617 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.851792 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.852368 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.853309 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.853623 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.854038 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.854003 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.854142 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.852101 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.854579 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.854863 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.855097 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.855302 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.856006 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.856230 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.857004 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.857222 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.857874 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.857941 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.858079 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.858529 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.858651 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.859026 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.859303 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.859517 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.859743 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.860092 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-kdjrf"] Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.860545 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:15:47.360522246 +0000 UTC m=+21.235667399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.861155 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.861416 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ftxxx"] Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.864135 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-k6hd5"] Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.865118 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.865760 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.865844 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866149 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866362 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866602 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866859 4831 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.867651 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866633 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.866922 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.867083 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.867241 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.867657 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.868639 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.868767 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.868951 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.868974 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.869151 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.869185 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.869340 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.869407 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.870207 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.871539 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.871948 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.872399 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.872418 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.872480 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.872801 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.872954 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.873583 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.873712 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.873793 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874118 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874335 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874536 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874686 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874806 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874820 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.875799 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.876035 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874915 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874946 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.876305 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.876447 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.874881 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.877044 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.878538 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.878705 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.879038 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.879042 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.879803 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.880266 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.880492 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.880723 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.880959 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.881089 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.881216 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.881351 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.881633 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.883104 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.883646 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.883971 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.888262 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.889446 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890104 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890448 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890592 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890617 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890720 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890853 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.890987 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.891649 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.892054 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.892289 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.894430 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.895194 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.896059 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.896421 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.896710 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.896844 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.896987 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.897235 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.897249 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.897467 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.899889 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.900686 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.902378 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.902644 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.902877 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.903136 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.903346 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.903491 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.903632 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.903976 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.904066 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.906482 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.906534 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.906734 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.906808 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.906866 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.907690 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.907784 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.908014 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.908604 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.913297 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.913649 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.913769 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.913851 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:47.413826616 +0000 UTC m=+21.288971959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.913938 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.913953 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.913963 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.914017 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:47.414002611 +0000 UTC m=+21.289147764 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.914057 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.914104 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:47.414092883 +0000 UTC m=+21.289238126 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.914275 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.915098 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.919536 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.932676 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933140 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht772\" (UniqueName: \"kubernetes.io/projected/f8de0f5f-48af-4be5-a67e-577d6f2897f9-kube-api-access-ht772\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933166 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-proxy-tls\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933185 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-binary-copy\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933416 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933456 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-hostroot\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933506 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-os-release\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933523 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-netns\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933537 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-conf-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933552 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-socket-dir-parent\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933576 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933593 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933638 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7ztt\" (UniqueName: \"kubernetes.io/projected/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-kube-api-access-q7ztt\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933678 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-k8s-cni-cncf-io\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933695 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-multus-daemon-config\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933718 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-bin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933748 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-etc-kubernetes\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933767 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933781 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-mcd-auth-proxy-config\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933825 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-multus\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933839 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933854 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-system-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933870 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-cni-binary-copy\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933905 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-cnibin\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933920 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-cnibin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933938 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-rootfs\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933954 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-system-cni-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933982 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.933987 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-multus-certs\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934037 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f8de0f5f-48af-4be5-a67e-577d6f2897f9-hosts-file\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934061 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934104 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jkfp\" (UniqueName: \"kubernetes.io/projected/66dfbe4f-ed46-4f54-b304-427a18c7da39-kube-api-access-5jkfp\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934153 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/f8de0f5f-48af-4be5-a67e-577d6f2897f9-hosts-file\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934191 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-os-release\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934223 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-kubelet\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934248 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj7wc\" (UniqueName: \"kubernetes.io/projected/89aea50f-21a1-4854-a107-6e38c780166d-kube-api-access-bj7wc\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934377 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934388 4831 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934397 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934421 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934429 4831 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934437 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934444 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934452 4831 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934514 4831 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934522 4831 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934530 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934537 4831 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934545 4831 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934578 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934596 4831 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934605 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934614 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934622 4831 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934629 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934652 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934661 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934669 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934676 4831 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934684 4831 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934693 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934700 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934723 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934731 4831 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934739 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934747 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934755 4831 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934762 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934769 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934777 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934785 4831 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934807 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934815 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934842 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934850 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934859 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934896 4831 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934905 4831 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934913 4831 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934921 4831 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934929 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.934938 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935354 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935369 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935379 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935387 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935396 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935404 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935428 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935437 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935444 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935453 4831 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935460 4831 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935468 4831 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935476 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935498 4831 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935506 4831 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935514 4831 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935522 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935529 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935539 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935547 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935555 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935576 4831 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935584 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935593 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935601 4831 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935609 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935616 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935625 4831 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935633 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935656 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935664 4831 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935672 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935680 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935688 4831 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935696 4831 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935704 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935712 4831 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935734 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935742 4831 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935750 4831 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935758 4831 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935766 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935774 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935781 4831 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935790 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935811 4831 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935818 4831 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935826 4831 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935834 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935842 4831 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935849 4831 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935864 4831 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935886 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935894 4831 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935901 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935909 4831 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935917 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935925 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935933 4831 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935940 4831 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935962 4831 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935970 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935979 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.935989 4831 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937004 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937020 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937031 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937042 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937050 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937077 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937086 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937095 4831 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937112 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937120 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937129 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937153 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937161 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937178 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937187 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937195 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937203 4831 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937212 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937236 4831 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937244 4831 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937252 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937260 4831 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937269 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937276 4831 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937285 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937306 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937324 4831 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937332 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937341 4831 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937350 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937390 4831 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937399 4831 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937408 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937421 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937430 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937440 4831 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937465 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937473 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937481 4831 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937489 4831 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937497 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937504 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937512 4831 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937520 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937544 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937553 4831 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937562 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937570 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937577 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937585 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937593 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937602 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937624 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937633 4831 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937643 4831 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937650 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937658 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937666 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937673 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937681 4831 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937704 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937712 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937719 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937727 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937734 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937744 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937752 4831 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937773 4831 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937781 4831 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937782 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937789 4831 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937825 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937834 4831 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.937602 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.945912 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.946279 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.949106 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.951401 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.957159 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht772\" (UniqueName: \"kubernetes.io/projected/f8de0f5f-48af-4be5-a67e-577d6f2897f9-kube-api-access-ht772\") pod \"node-resolver-556kb\" (UID: \"f8de0f5f-48af-4be5-a67e-577d6f2897f9\") " pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.957214 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.959033 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.959060 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.959072 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:46 crc kubenswrapper[4831]: E1124 08:15:46.959135 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:47.459103712 +0000 UTC m=+21.334248865 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.962460 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.962979 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.963726 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.966532 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.966628 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.967625 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.969831 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 24 08:15:46 crc kubenswrapper[4831]: I1124 08:15:46.991753 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.001573 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.002122 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.002794 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.004261 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.004979 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.006669 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.007190 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.008983 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.009619 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.013807 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.014235 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.025531 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.030752 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.035162 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.035698 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.036765 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.037218 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038144 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-os-release\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038181 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-kubelet\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038199 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj7wc\" (UniqueName: \"kubernetes.io/projected/89aea50f-21a1-4854-a107-6e38c780166d-kube-api-access-bj7wc\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038228 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-proxy-tls\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038245 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-binary-copy\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038260 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038277 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-hostroot\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038299 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-os-release\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038332 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-netns\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038363 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-conf-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038381 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-socket-dir-parent\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038401 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038425 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7ztt\" (UniqueName: \"kubernetes.io/projected/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-kube-api-access-q7ztt\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038443 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-k8s-cni-cncf-io\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038458 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-multus-daemon-config\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038475 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-bin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038493 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-etc-kubernetes\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038510 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-mcd-auth-proxy-config\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038524 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-multus\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038548 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038566 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-system-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038584 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-cni-binary-copy\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038601 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-cnibin\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038615 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-cnibin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038632 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-rootfs\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038648 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-system-cni-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038664 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-multus-certs\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038697 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jkfp\" (UniqueName: \"kubernetes.io/projected/66dfbe4f-ed46-4f54-b304-427a18c7da39-kube-api-access-5jkfp\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038730 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038742 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038755 4831 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038764 4831 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.038775 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.039090 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-os-release\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.039125 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-kubelet\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.039409 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-bin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040010 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-binary-copy\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040162 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-cni-binary-copy\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040170 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040195 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-hostroot\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040225 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-etc-kubernetes\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040247 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-os-release\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040278 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-netns\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040301 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-conf-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040351 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-multus-socket-dir-parent\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.040765 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-cnibin\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.041051 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.041458 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-tuning-conf-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.041551 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-cnibin\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.041582 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-rootfs\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.041819 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.042512 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-k8s-cni-cncf-io\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.042549 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-var-lib-cni-multus\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.042651 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/66dfbe4f-ed46-4f54-b304-427a18c7da39-system-cni-dir\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.042720 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-host-run-multus-certs\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.042961 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/66dfbe4f-ed46-4f54-b304-427a18c7da39-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.043033 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/89aea50f-21a1-4854-a107-6e38c780166d-system-cni-dir\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.043060 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/89aea50f-21a1-4854-a107-6e38c780166d-multus-daemon-config\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.043420 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.044461 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.045309 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.047572 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.047704 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.048256 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.049630 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-proxy-tls\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.061393 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-mcd-auth-proxy-config\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.061534 4831 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.061761 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.063784 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.064658 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.065628 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.067203 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.067960 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.069198 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.069920 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.070999 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.071905 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.073006 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jkfp\" (UniqueName: \"kubernetes.io/projected/66dfbe4f-ed46-4f54-b304-427a18c7da39-kube-api-access-5jkfp\") pod \"multus-additional-cni-plugins-kdjrf\" (UID: \"66dfbe4f-ed46-4f54-b304-427a18c7da39\") " pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.073046 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.074741 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7ztt\" (UniqueName: \"kubernetes.io/projected/8b857b1a-bb8a-4608-8643-e0a94ed82a1b-kube-api-access-q7ztt\") pod \"machine-config-daemon-ftxxx\" (UID: \"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\") " pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.076545 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj7wc\" (UniqueName: \"kubernetes.io/projected/89aea50f-21a1-4854-a107-6e38c780166d-kube-api-access-bj7wc\") pod \"multus-k6hd5\" (UID: \"89aea50f-21a1-4854-a107-6e38c780166d\") " pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.076974 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.077994 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.078608 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.078813 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.079888 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.081235 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.082508 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.083027 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.084100 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.084628 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.085797 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.086455 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.086946 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.096656 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.108985 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.118575 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.119448 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.128547 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.142099 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.153221 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.157524 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.163544 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.166897 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.178844 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: W1124 08:15:47.179633 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2f7be9a67e06865e284bb1483bf31cc6ad83117abdcc62e48cba7e9069be8ee5 WatchSource:0}: Error finding container 2f7be9a67e06865e284bb1483bf31cc6ad83117abdcc62e48cba7e9069be8ee5: Status 404 returned error can't find the container with id 2f7be9a67e06865e284bb1483bf31cc6ad83117abdcc62e48cba7e9069be8ee5 Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.186639 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-556kb" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.192250 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-k6hd5" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.220956 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.222515 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.235856 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hxqjw"] Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.236583 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.238796 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.239038 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.239169 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.239288 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.239409 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.239904 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.242840 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.248503 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.250038 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.266167 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.287018 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: W1124 08:15:47.303935 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66dfbe4f_ed46_4f54_b304_427a18c7da39.slice/crio-2cba396b01700ee0faaf7d92f764a38980d8df8d235cbf6c6420b0c5148a9f5c WatchSource:0}: Error finding container 2cba396b01700ee0faaf7d92f764a38980d8df8d235cbf6c6420b0c5148a9f5c: Status 404 returned error can't find the container with id 2cba396b01700ee0faaf7d92f764a38980d8df8d235cbf6c6420b0c5148a9f5c Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.311019 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.322792 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.339630 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345714 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345746 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345774 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345789 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345811 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345825 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345841 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345856 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345870 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345884 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345925 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345941 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345955 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345971 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.345985 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.346000 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.346026 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8gjg\" (UniqueName: \"kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.346053 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.346069 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.346084 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.356641 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.366856 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.389913 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.412277 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.441646 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.447766 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.447893 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.447922 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.447951 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:15:48.447934595 +0000 UTC m=+22.323079748 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.447969 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.447974 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448003 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448005 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448032 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448076 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448158 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448022 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448677 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448697 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448713 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448751 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448770 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448784 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448815 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8gjg\" (UniqueName: \"kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448834 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448854 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448871 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448888 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448903 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448919 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448942 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448958 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448974 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.448996 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.449034 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.449057 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.449077 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.449097 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450055 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450137 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450601 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450648 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450793 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450815 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450826 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450859 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:48.45084893 +0000 UTC m=+22.325994083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450886 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.450911 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450961 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.450983 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:48.450976184 +0000 UTC m=+22.326121337 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.451002 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.451022 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.451283 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.451366 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.451498 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.451635 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:48.451626082 +0000 UTC m=+22.326771235 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.466608 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.472522 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.473164 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8gjg\" (UniqueName: \"kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg\") pod \"ovnkube-node-hxqjw\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.482392 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.493217 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.515036 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.549477 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.549672 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.549715 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.549730 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:47 crc kubenswrapper[4831]: E1124 08:15:47.549789 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:48.549772816 +0000 UTC m=+22.424917969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:47 crc kubenswrapper[4831]: I1124 08:15:47.599390 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:47 crc kubenswrapper[4831]: W1124 08:15:47.618524 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77826b29_ac89_4b2a_afd9_746c0b1c13ff.slice/crio-b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c WatchSource:0}: Error finding container b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c: Status 404 returned error can't find the container with id b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.006998 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerStarted","Data":"976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.007048 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerStarted","Data":"e17851cc63ad6ce70d5e83c3ed11efe2a1fd8f34e6452c2e82c8892bf321afd3"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.009383 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-556kb" event={"ID":"f8de0f5f-48af-4be5-a67e-577d6f2897f9","Type":"ContainerStarted","Data":"39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.009423 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-556kb" event={"ID":"f8de0f5f-48af-4be5-a67e-577d6f2897f9","Type":"ContainerStarted","Data":"1b0d3936acc4a6709cfdd4c4335d97d685ee995a2dd9911f136f5c9c8bf6cb87"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.011339 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2f7be9a67e06865e284bb1483bf31cc6ad83117abdcc62e48cba7e9069be8ee5"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.013103 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.013137 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.013188 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"ed00d15887f9054a579299c80b6d05b8df4c8f3b434568f5b867dcff15bf7cb1"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.014843 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.014873 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3b5129fe7bdfd760fe720b8a91aeda6604207e5c5fef7113ab8204938609dc0b"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.019219 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.019269 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.019284 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"73d9c5131b681047eb8a3f57a1fa9bd6b3be0aa3d6feb487dc8f36df2dd47513"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.021132 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb" exitCode=0 Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.021189 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.021213 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.023108 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.023676 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.025480 4831 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117" exitCode=255 Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.025552 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.025592 4831 scope.go:117] "RemoveContainer" containerID="f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.028301 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23" exitCode=0 Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.028386 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.028410 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerStarted","Data":"2cba396b01700ee0faaf7d92f764a38980d8df8d235cbf6c6420b0c5148a9f5c"} Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.038026 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.064242 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.064635 4831 scope.go:117] "RemoveContainer" containerID="589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.064712 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.064825 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.093480 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.106508 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.127735 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.147872 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.157917 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.183738 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.202555 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.215447 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.234768 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.253718 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.269011 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.280655 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.297111 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.316943 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.340738 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f64f03b976f902b782da65508e8ecd62c7897636221c3d021dc84eef5ed12351\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:40Z\\\",\\\"message\\\":\\\"W1124 08:15:30.053418 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1124 08:15:30.053712 1 crypto.go:601] Generating new CA for check-endpoints-signer@1763972130 cert, and key in /tmp/serving-cert-487117384/serving-signer.crt, /tmp/serving-cert-487117384/serving-signer.key\\\\nI1124 08:15:30.496802 1 observer_polling.go:159] Starting file observer\\\\nW1124 08:15:30.499645 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1124 08:15:30.499848 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:30.502525 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-487117384/tls.crt::/tmp/serving-cert-487117384/tls.key\\\\\\\"\\\\nF1124 08:15:40.836635 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.354082 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.370681 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.385299 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.405430 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.418111 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.430444 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:48Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.458980 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.459055 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.459079 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459165 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459155 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:15:50.459128825 +0000 UTC m=+24.334273978 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459247 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:50.459240138 +0000 UTC m=+24.334385291 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459267 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459307 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459342 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.459356 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459405 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:50.459382222 +0000 UTC m=+24.334527435 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459450 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.459499 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:50.459492396 +0000 UTC m=+24.334637549 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.560636 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.560827 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.560852 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.560905 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.560969 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:50.560951495 +0000 UTC m=+24.436096648 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.893183 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.893204 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.893340 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.893718 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.893791 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:48 crc kubenswrapper[4831]: E1124 08:15:48.893858 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.896704 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 24 08:15:48 crc kubenswrapper[4831]: I1124 08:15:48.897656 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.033197 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.035577 4831 scope.go:117] "RemoveContainer" containerID="589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117" Nov 24 08:15:49 crc kubenswrapper[4831]: E1124 08:15:49.035785 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.045682 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.045722 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.045755 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.045765 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.045773 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.047343 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerStarted","Data":"b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f"} Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.055587 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.067292 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-mnrjr"] Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.067709 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.069327 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.069597 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.069830 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.070001 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.071249 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.081102 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.096950 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.113170 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.126866 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.138356 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.151542 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.162868 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.165397 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khwh7\" (UniqueName: \"kubernetes.io/projected/9da0997c-8ea4-4af7-b1db-428d24f7d635-kube-api-access-khwh7\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.165523 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9da0997c-8ea4-4af7-b1db-428d24f7d635-serviceca\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.165595 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da0997c-8ea4-4af7-b1db-428d24f7d635-host\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.176010 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.193256 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.193237 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.197422 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.200733 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.218807 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.231341 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.244987 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.257596 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.266386 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khwh7\" (UniqueName: \"kubernetes.io/projected/9da0997c-8ea4-4af7-b1db-428d24f7d635-kube-api-access-khwh7\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.266429 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9da0997c-8ea4-4af7-b1db-428d24f7d635-serviceca\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.266457 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da0997c-8ea4-4af7-b1db-428d24f7d635-host\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.266580 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9da0997c-8ea4-4af7-b1db-428d24f7d635-host\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.267841 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9da0997c-8ea4-4af7-b1db-428d24f7d635-serviceca\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.271274 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.284929 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.296068 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.314857 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.331290 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.342755 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.352044 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khwh7\" (UniqueName: \"kubernetes.io/projected/9da0997c-8ea4-4af7-b1db-428d24f7d635-kube-api-access-khwh7\") pod \"node-ca-mnrjr\" (UID: \"9da0997c-8ea4-4af7-b1db-428d24f7d635\") " pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.359346 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.376176 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.391838 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.407928 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-mnrjr" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.426139 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.443636 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.516048 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.634468 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.651358 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.655375 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.672582 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.692784 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.708429 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.721550 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.733340 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.742799 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.753953 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.764331 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.780501 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.792296 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.807486 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.816715 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.836226 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.854290 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.879690 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.911788 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.929705 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.945227 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:49 crc kubenswrapper[4831]: I1124 08:15:49.971603 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:49Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.014480 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.051662 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-mnrjr" event={"ID":"9da0997c-8ea4-4af7-b1db-428d24f7d635","Type":"ContainerStarted","Data":"86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107"} Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.051888 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-mnrjr" event={"ID":"9da0997c-8ea4-4af7-b1db-428d24f7d635","Type":"ContainerStarted","Data":"9575ee9702649cf115d9ab2a5ec1e0087f2872735383d6e4c331b76ace3f1445"} Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.055579 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2"} Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.057415 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f" exitCode=0 Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.057483 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f"} Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.059825 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b"} Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.060916 4831 scope.go:117] "RemoveContainer" containerID="589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.061051 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.071854 4831 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.074156 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.110812 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.152573 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.192521 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.231005 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.272400 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.311765 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.356188 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.392712 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.432298 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.471421 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.478782 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.478970 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:15:54.478934331 +0000 UTC m=+28.354079524 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.479302 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.479637 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.479812 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.479461 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.480104 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:54.480090515 +0000 UTC m=+28.355235678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.479843 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.480740 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.479949 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.480948 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:54.480918519 +0000 UTC m=+28.356063712 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.480856 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.481313 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:54.48129194 +0000 UTC m=+28.356437113 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.515650 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.554537 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.580413 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.580584 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.580605 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.580619 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.580669 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:15:54.580651728 +0000 UTC m=+28.455796881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.593498 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.632712 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.676025 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.711945 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.749173 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.795503 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.830433 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.871441 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.892566 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.892672 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.892703 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.892771 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.892853 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:50 crc kubenswrapper[4831]: E1124 08:15:50.893031 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.910058 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.951951 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:50 crc kubenswrapper[4831]: I1124 08:15:50.995139 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:50Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.064505 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9" exitCode=0 Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.064567 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9"} Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.087202 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.113230 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.127030 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.151642 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.190722 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.230561 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.269868 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.311025 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.352423 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.391517 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.431705 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.473513 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.510300 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.552806 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:51 crc kubenswrapper[4831]: I1124 08:15:51.590290 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:51Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.070596 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8"} Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.072691 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa" exitCode=0 Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.072765 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa"} Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.089773 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.101416 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.112684 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.123836 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.136181 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.149405 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.160754 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.170836 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.189401 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.208018 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.217803 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.230396 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.244081 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.256015 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.265279 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.518644 4831 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.521160 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.521207 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.521217 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.521348 4831 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.528280 4831 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.528731 4831 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.529870 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.529906 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.529916 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.529934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.529944 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.541981 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.545718 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.545753 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.545765 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.545781 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.545790 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.556914 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.560913 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.560953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.560961 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.560978 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.560989 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.574597 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.578934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.579022 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.579040 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.579067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.579081 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.592447 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.600087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.600146 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.600159 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.600180 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.600226 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.616090 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:52Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.616212 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.617863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.617895 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.617903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.617915 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.617925 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.720980 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.721032 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.721046 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.721070 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.721087 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.822999 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.823035 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.823046 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.823063 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.823075 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.893273 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.893339 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.893276 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.893440 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.893491 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:52 crc kubenswrapper[4831]: E1124 08:15:52.893538 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.925957 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.925997 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.926006 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.926020 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:52 crc kubenswrapper[4831]: I1124 08:15:52.926029 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:52Z","lastTransitionTime":"2025-11-24T08:15:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.028947 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.028982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.028993 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.029008 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.029020 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.079616 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e" exitCode=0 Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.079656 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.106782 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.125163 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.131312 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.131374 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.131388 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.131404 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.131415 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.141097 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.164532 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.178685 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.190380 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.208936 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.226380 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.234036 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.234082 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.234096 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.234118 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.234132 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.243960 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.262662 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.314256 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.328892 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.335845 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.335885 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.335896 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.335912 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.335924 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.342039 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.354613 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.368973 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:53Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.439419 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.439467 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.439477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.439494 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.439509 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.542409 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.542452 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.542464 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.542480 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.542492 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.650296 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.650512 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.650521 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.650535 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.650544 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.753129 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.753158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.753168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.753181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.753190 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.855111 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.855156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.855165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.855180 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.855188 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.958334 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.958383 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.958393 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.958411 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:53 crc kubenswrapper[4831]: I1124 08:15:53.958422 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:53Z","lastTransitionTime":"2025-11-24T08:15:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.061902 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.061955 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.061968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.061989 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.062003 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.087674 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.088029 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.091576 4831 generic.go:334] "Generic (PLEG): container finished" podID="66dfbe4f-ed46-4f54-b304-427a18c7da39" containerID="82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404" exitCode=0 Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.091618 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerDied","Data":"82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.112892 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.117456 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.128080 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.140604 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.159199 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.164675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.164720 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.164734 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.164759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.164772 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.177899 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.193846 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.208241 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.247121 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.264563 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.267185 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.267218 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.267228 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.267244 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.267257 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.283646 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.297258 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.310952 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.324649 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.338664 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.350226 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.361754 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.373926 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.374783 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.374798 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.374824 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.374837 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.379250 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.395756 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.410635 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.424795 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.440601 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.458518 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.474728 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.477790 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.477835 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.477850 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.477867 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.477878 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.491518 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.505519 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.519975 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.537161 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.537375 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537426 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.537390871 +0000 UTC m=+36.412536024 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537488 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537545 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.537529286 +0000 UTC m=+36.412674439 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537633 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537656 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537674 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537714 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.537704031 +0000 UTC m=+36.412849254 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.537487 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.537776 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537864 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.537896 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.537887806 +0000 UTC m=+36.413032959 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.577063 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.580875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.580918 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.580928 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.580950 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.580962 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.592654 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.615025 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.634111 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:54Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.638456 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.638602 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.638619 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.638632 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.638686 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.638669586 +0000 UTC m=+36.513814739 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.683485 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.683525 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.683533 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.683549 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.683565 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.786872 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.786921 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.786958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.786982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.786999 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.890498 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.890556 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.890571 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.890593 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.890608 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.893172 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.893430 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.893439 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.893514 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.893527 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:54 crc kubenswrapper[4831]: E1124 08:15:54.893686 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.993748 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.993799 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.993811 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.993829 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:54 crc kubenswrapper[4831]: I1124 08:15:54.993843 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:54Z","lastTransitionTime":"2025-11-24T08:15:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.095605 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.095634 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.095644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.095660 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.095673 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.099151 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.099813 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" event={"ID":"66dfbe4f-ed46-4f54-b304-427a18c7da39","Type":"ContainerStarted","Data":"f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.099889 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.115161 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.127100 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.127972 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.140621 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.153666 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.164828 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.181821 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.195281 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.197819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.197854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.197863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.197882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.197897 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.244129 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.281255 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.300339 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.300380 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.300388 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.300408 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.300417 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.302557 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.321744 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.334428 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.347246 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.360875 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.379598 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.403819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.403882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.403893 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.403914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.403925 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.409059 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.421832 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.434528 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.454309 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.470243 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.484545 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.495141 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.506112 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.506158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.506170 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.506186 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.506196 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.510788 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.526108 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.542060 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.554800 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.573478 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.587673 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.604333 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.608399 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.608431 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.608440 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.608455 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.608465 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.616236 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:55Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.710837 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.710886 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.710899 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.710918 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.710936 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.812851 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.812881 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.812889 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.812903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.812912 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.914586 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.914628 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.914651 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.914668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:55 crc kubenswrapper[4831]: I1124 08:15:55.914679 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:55Z","lastTransitionTime":"2025-11-24T08:15:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.016722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.016746 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.016754 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.016766 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.016776 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.101358 4831 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.120723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.120759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.120768 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.120781 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.120789 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.223249 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.223290 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.223301 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.223342 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.223358 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.325870 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.325909 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.325918 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.325932 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.325942 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.427972 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.428004 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.428012 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.428026 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.428035 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.430126 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.530824 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.530862 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.530873 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.530890 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.530902 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.633637 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.633701 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.633733 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.633758 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.633786 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.736003 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.736034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.736042 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.736057 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.736066 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.837933 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.837991 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.838001 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.838014 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.838024 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.892462 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.892598 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.892803 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:56 crc kubenswrapper[4831]: E1124 08:15:56.892791 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:56 crc kubenswrapper[4831]: E1124 08:15:56.892918 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:56 crc kubenswrapper[4831]: E1124 08:15:56.892997 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.912876 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.922691 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.935256 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.940052 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.940092 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.940101 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.940116 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.940125 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:56Z","lastTransitionTime":"2025-11-24T08:15:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.949053 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.960335 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.973594 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:56 crc kubenswrapper[4831]: I1124 08:15:56.989032 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:56Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.011697 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.031191 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.045234 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.045475 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.045601 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.045824 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.045916 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.051035 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.067356 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.083077 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.098964 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.105467 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/0.log" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.107931 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3" exitCode=1 Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.108064 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.108673 4831 scope.go:117] "RemoveContainer" containerID="d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.109941 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.126368 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.137518 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.148493 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.149745 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.149773 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.149784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.149802 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.149814 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.163371 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.175006 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.189682 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.202281 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.221615 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:56Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564242 6022 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:15:56.564266 6022 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:56.564282 6022 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:56.564298 6022 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:56.564345 6022 factory.go:656] Stopping watch factory\\\\nI1124 08:15:56.564366 6022 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:56.564387 6022 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:56.564403 6022 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:56.564449 6022 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564590 6022 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564620 6022 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.252106 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.252247 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.252422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.252487 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.252599 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.257916 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.270339 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.287045 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.300721 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.312724 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.322978 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.336775 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.349564 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:57Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.356018 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.356055 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.356063 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.356079 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.356089 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.458645 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.458693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.458705 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.458723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.458735 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.561059 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.561105 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.561118 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.561134 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.561144 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.663812 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.663841 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.663850 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.663863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.663872 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.766045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.766108 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.766127 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.766151 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.766166 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.868671 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.868711 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.868722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.868741 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.868751 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.972165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.972210 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.972222 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.972237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:57 crc kubenswrapper[4831]: I1124 08:15:57.972248 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:57Z","lastTransitionTime":"2025-11-24T08:15:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.074612 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.074639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.074647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.074689 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.074699 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.112691 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/1.log" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.113260 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/0.log" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.115284 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a" exitCode=1 Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.115342 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.115377 4831 scope.go:117] "RemoveContainer" containerID="d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.115922 4831 scope.go:117] "RemoveContainer" containerID="4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a" Nov 24 08:15:58 crc kubenswrapper[4831]: E1124 08:15:58.116051 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.131052 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.143617 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.154897 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.166578 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.176486 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.176521 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.176532 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.176548 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.176567 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.180220 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.190957 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.209031 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6451db44192875f47fd2f32d880318bd1d17bda92e05d9730f562dacf3f05b3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:56Z\\\",\\\"message\\\":\\\"go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564242 6022 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1124 08:15:56.564266 6022 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:56.564282 6022 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:56.564298 6022 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:56.564345 6022 factory.go:656] Stopping watch factory\\\\nI1124 08:15:56.564366 6022 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:56.564387 6022 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:56.564403 6022 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:56.564449 6022 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564590 6022 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1124 08:15:56.564620 6022 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.230575 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.243472 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.253972 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.265061 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278150 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278606 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278628 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278637 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278649 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.278657 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.287452 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.299231 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.312223 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:58Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.381030 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.381062 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.381070 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.381087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.381095 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.483251 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.483286 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.483296 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.483329 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.483341 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.585986 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.586016 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.586026 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.586041 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.586052 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.688117 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.688455 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.688465 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.688481 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.688493 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.789890 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.789917 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.789926 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.789939 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.789949 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.891826 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.891851 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.891859 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.891871 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.891882 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.892359 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.892377 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.892377 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:15:58 crc kubenswrapper[4831]: E1124 08:15:58.892452 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:15:58 crc kubenswrapper[4831]: E1124 08:15:58.892536 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:15:58 crc kubenswrapper[4831]: E1124 08:15:58.892625 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.994015 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.994048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.994057 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.994070 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:58 crc kubenswrapper[4831]: I1124 08:15:58.994079 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:58Z","lastTransitionTime":"2025-11-24T08:15:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.096654 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.096693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.096704 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.096720 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.096729 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.119192 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/1.log" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.122971 4831 scope.go:117] "RemoveContainer" containerID="4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a" Nov 24 08:15:59 crc kubenswrapper[4831]: E1124 08:15:59.123214 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.134874 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.155131 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.173724 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.185980 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.198791 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.198827 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.198837 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.198850 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.198859 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.199481 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.212010 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.222729 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.231965 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.243297 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.255870 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.266670 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.278906 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.292229 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.300961 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.301002 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.301013 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.301030 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.301039 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.306203 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.318652 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.402888 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.402946 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.402958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.402975 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.402985 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.505239 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.505286 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.505341 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.505362 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.505374 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.607958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.608005 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.608013 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.608052 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.608065 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.709575 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.709790 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.709981 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.710178 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.710385 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.812184 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.812227 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.812239 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.812255 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.812265 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.869951 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft"] Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.870405 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.872650 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.873894 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.888026 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.898457 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.907098 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.914888 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.914922 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.914933 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.914949 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.914960 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:15:59Z","lastTransitionTime":"2025-11-24T08:15:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.918162 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.930739 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.943407 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.959498 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.971069 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.983217 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.989819 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bjbt\" (UniqueName: \"kubernetes.io/projected/aa82b248-7cf0-4285-8978-9ef7d10a265b-kube-api-access-9bjbt\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.989885 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.989948 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.989972 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:15:59 crc kubenswrapper[4831]: I1124 08:15:59.998730 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:15:59Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.010437 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.017105 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.017134 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.017144 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.017158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.017166 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.020103 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.037242 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.048232 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.058641 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.074738 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.090482 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bjbt\" (UniqueName: \"kubernetes.io/projected/aa82b248-7cf0-4285-8978-9ef7d10a265b-kube-api-access-9bjbt\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.090546 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.090581 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.090598 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.091118 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.091369 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.098078 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aa82b248-7cf0-4285-8978-9ef7d10a265b-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.106896 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bjbt\" (UniqueName: \"kubernetes.io/projected/aa82b248-7cf0-4285-8978-9ef7d10a265b-kube-api-access-9bjbt\") pod \"ovnkube-control-plane-749d76644c-2dnft\" (UID: \"aa82b248-7cf0-4285-8978-9ef7d10a265b\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.119193 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.119237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.119245 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.119258 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.119268 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.182461 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" Nov 24 08:16:00 crc kubenswrapper[4831]: W1124 08:16:00.194411 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa82b248_7cf0_4285_8978_9ef7d10a265b.slice/crio-fce3603c47f39348dcaf99258b8a7760cfe1e4b7b982557934b662b3fb39755b WatchSource:0}: Error finding container fce3603c47f39348dcaf99258b8a7760cfe1e4b7b982557934b662b3fb39755b: Status 404 returned error can't find the container with id fce3603c47f39348dcaf99258b8a7760cfe1e4b7b982557934b662b3fb39755b Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.221299 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.221380 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.221393 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.221409 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.221421 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.324168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.324210 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.324220 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.324234 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.324243 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.426815 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.426845 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.426854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.426871 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.426880 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.528676 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.528711 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.528721 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.528736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.528748 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.590526 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-k45nx"] Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.591026 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.591095 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.603686 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.613559 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.624852 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.631134 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.631176 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.631187 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.631203 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.631213 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.633751 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.642827 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.654507 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.664614 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.674930 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.684536 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.695220 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.695261 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh6dz\" (UniqueName: \"kubernetes.io/projected/650855b7-20e6-4261-a212-4bc75a671b96-kube-api-access-mh6dz\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.700079 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.716095 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.724621 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.733803 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.733837 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.733868 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.733882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.733892 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.736345 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.748283 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.757121 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.768085 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.778096 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:00Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.796780 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.796819 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh6dz\" (UniqueName: \"kubernetes.io/projected/650855b7-20e6-4261-a212-4bc75a671b96-kube-api-access-mh6dz\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.796983 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.797088 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:01.297055942 +0000 UTC m=+35.172201095 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.811279 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh6dz\" (UniqueName: \"kubernetes.io/projected/650855b7-20e6-4261-a212-4bc75a671b96-kube-api-access-mh6dz\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.838599 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.838639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.838651 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.838670 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.838687 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.892411 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.892512 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.892587 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.892621 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.892731 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:00 crc kubenswrapper[4831]: E1124 08:16:00.893055 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.941339 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.941387 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.941396 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.941410 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:00 crc kubenswrapper[4831]: I1124 08:16:00.941420 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:00Z","lastTransitionTime":"2025-11-24T08:16:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.043587 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.043622 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.043632 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.043647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.043657 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.130003 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" event={"ID":"aa82b248-7cf0-4285-8978-9ef7d10a265b","Type":"ContainerStarted","Data":"c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.130059 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" event={"ID":"aa82b248-7cf0-4285-8978-9ef7d10a265b","Type":"ContainerStarted","Data":"2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.130073 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" event={"ID":"aa82b248-7cf0-4285-8978-9ef7d10a265b","Type":"ContainerStarted","Data":"fce3603c47f39348dcaf99258b8a7760cfe1e4b7b982557934b662b3fb39755b"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.145843 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.145879 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.145889 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.145904 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.145915 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.149131 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.159949 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.170404 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.186212 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.199199 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.210528 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.221556 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.233162 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.245299 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.247784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.247815 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.247823 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.247835 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.247843 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.254813 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.266257 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.277555 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.288891 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.300409 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.300813 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:01 crc kubenswrapper[4831]: E1124 08:16:01.300953 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:01 crc kubenswrapper[4831]: E1124 08:16:01.301047 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:02.300986114 +0000 UTC m=+36.176131267 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.327807 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349214 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349846 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349873 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349883 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349898 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.349909 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.375217 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:01Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.452139 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.452171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.452181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.452198 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.452208 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.555415 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.555493 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.555534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.555557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.555573 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.657925 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.657956 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.657964 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.657978 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.657986 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.760747 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.760792 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.760808 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.760828 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.760842 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.863101 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.863146 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.863161 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.863183 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.863195 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.965468 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.965500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.965512 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.965536 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:01 crc kubenswrapper[4831]: I1124 08:16:01.965548 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:01Z","lastTransitionTime":"2025-11-24T08:16:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.067602 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.067644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.067654 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.067671 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.067682 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.173028 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.173067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.173076 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.173094 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.173105 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.278220 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.278251 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.278259 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.278273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.278284 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.311769 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.311957 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.312045 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:04.31202452 +0000 UTC m=+38.187169673 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.381041 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.381114 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.381131 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.381156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.381170 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.483048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.483089 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.483097 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.483110 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.483119 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.585989 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.586031 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.586043 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.586062 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.586077 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.614777 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.615066 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.615102 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615139 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:16:18.615099561 +0000 UTC m=+52.490244704 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.615212 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615235 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615299 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:18.615287287 +0000 UTC m=+52.490432630 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615310 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615362 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615383 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615469 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:18.615449691 +0000 UTC m=+52.490594844 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615539 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.615734 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:18.615662788 +0000 UTC m=+52.490808101 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.674704 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.674776 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.674785 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.675015 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.675027 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.689066 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.694056 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.694130 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.694177 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.694199 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.694214 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.707774 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.712082 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.712126 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.712137 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.712154 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.712166 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.715983 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.716109 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.716132 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.716142 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.716186 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:18.71617388 +0000 UTC m=+52.591319033 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.730435 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.735103 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.735139 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.735149 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.735163 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.735174 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.749033 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.752803 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.752836 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.752848 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.752865 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.752873 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.767115 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:02Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.767234 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.768646 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.768669 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.768696 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.768712 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.768723 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.871457 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.871492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.871503 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.871519 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.871531 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.892826 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.892975 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.893464 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.893537 4831 scope.go:117] "RemoveContainer" containerID="589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.893667 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.893845 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.893894 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.894135 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:02 crc kubenswrapper[4831]: E1124 08:16:02.894231 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.973892 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.973924 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.973932 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.973945 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:02 crc kubenswrapper[4831]: I1124 08:16:02.973980 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:02Z","lastTransitionTime":"2025-11-24T08:16:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.075897 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.076262 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.076271 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.076304 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.076333 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.139984 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.141704 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.142255 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.167192 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.178354 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.178390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.178401 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.178415 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.178424 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.192877 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.206431 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.225662 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.236734 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.249934 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.261345 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.272269 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.280981 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.281008 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.281016 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.281045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.281054 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.285582 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.297342 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.316876 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.328622 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.340100 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.352673 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.365147 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.375859 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.382920 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.382957 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.382968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.382984 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.382993 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.385296 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:03Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.485463 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.485562 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.485572 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.485585 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.485603 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.588553 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.588902 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.588988 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.589076 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.589151 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.691530 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.691794 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.691910 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.692009 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.692109 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.794557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.794609 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.794619 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.794635 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.794653 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.896748 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.896976 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.897079 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.897170 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:03 crc kubenswrapper[4831]: I1124 08:16:03.897292 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:03Z","lastTransitionTime":"2025-11-24T08:16:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.000489 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.000732 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.000849 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.000973 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.001045 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.103834 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.104127 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.104192 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.104276 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.104402 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.206484 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.206523 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.206532 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.206545 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.206557 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.309117 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.309149 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.309159 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.309184 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.309193 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.331024 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.331188 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.331409 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:08.331394183 +0000 UTC m=+42.206539336 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.410747 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.410828 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.410838 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.410854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.410866 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.513759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.513809 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.513823 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.513845 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.513855 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.616723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.616774 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.616792 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.616819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.616838 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.719522 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.719567 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.719581 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.719604 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.719623 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.822125 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.822156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.822164 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.822176 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.822185 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.892787 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.893062 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.892801 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.893157 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.893183 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.893258 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.893345 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:04 crc kubenswrapper[4831]: E1124 08:16:04.893391 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.924547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.924620 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.924632 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.924672 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:04 crc kubenswrapper[4831]: I1124 08:16:04.924686 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:04Z","lastTransitionTime":"2025-11-24T08:16:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.027198 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.027237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.027247 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.027260 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.027270 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.130830 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.131156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.131293 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.131435 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.131514 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.234161 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.234220 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.234236 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.234263 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.234277 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.337278 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.337349 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.337359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.337372 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.337384 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.440236 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.440277 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.440288 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.440304 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.440314 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.543088 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.543143 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.543161 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.543177 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.543188 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.645279 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.645329 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.645338 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.645352 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.645360 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.747303 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.747354 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.747363 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.747377 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.747387 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.849132 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.849173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.849189 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.849205 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.849215 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.951385 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.951457 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.951467 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.951482 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:05 crc kubenswrapper[4831]: I1124 08:16:05.951493 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:05Z","lastTransitionTime":"2025-11-24T08:16:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.053169 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.053215 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.053226 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.053241 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.053252 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.154722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.154753 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.154762 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.154776 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.154785 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.256978 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.257011 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.257020 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.257034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.257043 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.358723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.358769 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.358779 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.358797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.358808 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.460942 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.460974 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.460982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.460995 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.461008 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.563214 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.563256 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.563266 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.563282 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.563292 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.665859 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.665895 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.665903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.665917 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.665925 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.767882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.767910 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.767918 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.767931 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.767939 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.870479 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.870510 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.870522 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.870537 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.870546 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.893238 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.893293 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.893370 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.893238 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:06 crc kubenswrapper[4831]: E1124 08:16:06.893426 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:06 crc kubenswrapper[4831]: E1124 08:16:06.893502 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:06 crc kubenswrapper[4831]: E1124 08:16:06.893561 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:06 crc kubenswrapper[4831]: E1124 08:16:06.893621 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.912481 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.929386 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.952256 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.973853 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.973906 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.973919 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.973939 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.973952 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:06Z","lastTransitionTime":"2025-11-24T08:16:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.977165 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:06 crc kubenswrapper[4831]: I1124 08:16:06.991279 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:06Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.009765 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.020926 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.033449 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.045772 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.059777 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.076928 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.076975 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.076987 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.077003 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.077015 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.077851 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.091571 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.102842 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.118157 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.136238 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.150571 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.163120 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:07Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.179699 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.179737 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.179746 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.179761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.179770 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.282347 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.282383 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.282394 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.282410 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.282422 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.385291 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.385403 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.385431 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.385464 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.385487 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.488953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.489024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.489045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.489073 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.489099 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.591417 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.591450 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.591461 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.591478 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.591488 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.693516 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.693553 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.693562 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.693575 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.693584 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.795723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.795761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.795773 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.795789 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.795801 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.902397 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.902445 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.902456 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.902473 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:07 crc kubenswrapper[4831]: I1124 08:16:07.902484 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:07Z","lastTransitionTime":"2025-11-24T08:16:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.004842 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.004885 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.004894 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.004908 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.004916 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.107419 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.107452 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.107463 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.107479 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.107490 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.210135 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.210166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.210176 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.210191 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.210202 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.311805 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.311838 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.311848 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.311864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.311875 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.372575 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.372718 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.372808 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:16.372789677 +0000 UTC m=+50.247934830 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.414607 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.414639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.414650 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.414665 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.414674 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.517024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.517133 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.517146 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.517164 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.517177 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.619604 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.619645 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.619656 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.619673 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.619684 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.722209 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.722248 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.722260 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.722279 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.722293 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.824656 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.824725 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.824736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.824750 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.824758 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.892899 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.892956 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.892930 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.892901 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.893063 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.893137 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.893240 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:08 crc kubenswrapper[4831]: E1124 08:16:08.893406 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.928531 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.928575 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.928588 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.928605 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:08 crc kubenswrapper[4831]: I1124 08:16:08.928616 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:08Z","lastTransitionTime":"2025-11-24T08:16:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.030437 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.030471 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.030481 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.030496 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.030508 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.132864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.132905 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.132914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.132929 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.132937 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.235589 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.235628 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.235638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.235656 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.235667 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.338239 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.338300 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.338395 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.338427 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.338444 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.441729 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.441793 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.441810 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.441837 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.441854 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.544500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.544538 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.544547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.544563 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.544571 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.647217 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.647265 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.647277 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.647290 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.647298 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.750096 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.750157 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.750166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.750182 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.750191 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.852932 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.852958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.852969 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.853001 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.853010 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.954937 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.954968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.954976 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.954989 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:09 crc kubenswrapper[4831]: I1124 08:16:09.955011 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:09Z","lastTransitionTime":"2025-11-24T08:16:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.057510 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.057786 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.057882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.058006 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.058085 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.160595 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.160631 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.160639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.160657 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.160668 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.264569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.264638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.264651 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.264668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.264680 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.367431 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.367482 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.367492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.367505 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.367516 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.469415 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.469455 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.469468 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.469491 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.469520 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.572487 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.572531 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.572539 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.572553 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.572562 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.676379 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.676417 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.676456 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.676472 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.676484 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.778704 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.778745 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.778756 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.778772 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.778783 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.881248 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.881286 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.881296 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.881310 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.881348 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.892582 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.892633 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.892576 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:10 crc kubenswrapper[4831]: E1124 08:16:10.892683 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:10 crc kubenswrapper[4831]: E1124 08:16:10.892812 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.892809 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:10 crc kubenswrapper[4831]: E1124 08:16:10.892968 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:10 crc kubenswrapper[4831]: E1124 08:16:10.893088 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.984778 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.984830 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.984846 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.984868 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:10 crc kubenswrapper[4831]: I1124 08:16:10.984884 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:10Z","lastTransitionTime":"2025-11-24T08:16:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.088122 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.088203 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.088221 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.088253 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.088272 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.190802 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.190842 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.190851 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.190864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.190875 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.293924 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.294287 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.294405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.294520 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.294717 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.397600 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.397631 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.397639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.397650 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.397659 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.500204 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.500241 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.500267 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.500281 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.500290 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.602668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.602699 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.602710 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.602726 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.602736 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.705746 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.705820 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.705835 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.705861 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.705876 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.808981 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.809058 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.809076 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.809102 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.809118 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.912461 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.912544 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.912560 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.912592 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:11 crc kubenswrapper[4831]: I1124 08:16:11.912610 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:11Z","lastTransitionTime":"2025-11-24T08:16:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.015917 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.016043 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.016053 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.016069 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.016079 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.118365 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.118401 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.118411 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.118426 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.118437 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.221642 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.221673 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.221681 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.221694 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.221702 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.325119 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.325190 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.325202 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.325223 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.325266 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.429163 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.429235 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.429250 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.429273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.429287 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.532782 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.532920 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.532990 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.533065 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.533090 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.635937 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.635995 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.636010 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.636034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.636053 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.738694 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.738739 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.738751 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.738770 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.738782 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.841561 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.841609 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.841620 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.841637 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.841648 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.893230 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.893343 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.893346 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:12 crc kubenswrapper[4831]: E1124 08:16:12.893420 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:12 crc kubenswrapper[4831]: E1124 08:16:12.893538 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.893829 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:12 crc kubenswrapper[4831]: E1124 08:16:12.893950 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:12 crc kubenswrapper[4831]: E1124 08:16:12.894016 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.944008 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.944055 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.944067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.944085 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.944097 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.970396 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.970438 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.970448 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.970465 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.970476 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:12 crc kubenswrapper[4831]: E1124 08:16:12.983603 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.987300 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.987578 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.987704 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.987798 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:12 crc kubenswrapper[4831]: I1124 08:16:12.987891 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:12Z","lastTransitionTime":"2025-11-24T08:16:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: E1124 08:16:13.001143 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:12Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.005424 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.005467 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.005476 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.005492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.005505 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: E1124 08:16:13.019905 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.023379 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.023422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.023436 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.023454 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.023466 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: E1124 08:16:13.036846 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.040398 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.040431 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.040441 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.040459 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.040470 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: E1124 08:16:13.050805 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:13Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:13 crc kubenswrapper[4831]: E1124 08:16:13.050934 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.052306 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.052348 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.052359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.052372 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.052379 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.154617 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.154645 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.154655 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.154671 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.154681 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.256724 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.256767 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.256775 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.256788 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.256797 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.359545 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.359586 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.359594 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.359608 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.359617 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.461495 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.461517 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.461525 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.461537 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.461546 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.564473 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.564519 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.564534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.564555 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.564571 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.666824 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.666881 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.666891 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.666906 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.666917 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.769217 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.769262 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.769281 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.769297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.769310 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.872226 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.872287 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.872300 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.872336 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.872347 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.893479 4831 scope.go:117] "RemoveContainer" containerID="4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.974258 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.974568 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.974580 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.974599 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:13 crc kubenswrapper[4831]: I1124 08:16:13.974611 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:13Z","lastTransitionTime":"2025-11-24T08:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.076733 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.077533 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.077742 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.077861 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.078021 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.177775 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/1.log" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.180002 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.180054 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.180067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.180090 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.180107 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.182512 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.187854 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.206261 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.227981 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.247519 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.264829 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.283342 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.283381 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.283391 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.283408 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.283419 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.287331 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.309631 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.327039 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.345300 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.360043 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.381433 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.385446 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.385470 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.385477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.385490 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.385501 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.406972 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.419840 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.434683 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.452154 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.468106 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.483978 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.487973 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.488027 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.488040 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.488060 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.488072 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.499756 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:14Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.591484 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.591560 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.591584 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.591615 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.591636 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.694034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.694089 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.694102 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.694122 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.694135 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.796976 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.797022 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.797034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.797053 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.797067 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.892457 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.892540 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.892536 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:14 crc kubenswrapper[4831]: E1124 08:16:14.892647 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.892718 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:14 crc kubenswrapper[4831]: E1124 08:16:14.892799 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:14 crc kubenswrapper[4831]: E1124 08:16:14.893168 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:14 crc kubenswrapper[4831]: E1124 08:16:14.893258 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.899130 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.899173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.899183 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.899200 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:14 crc kubenswrapper[4831]: I1124 08:16:14.899216 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:14Z","lastTransitionTime":"2025-11-24T08:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.001308 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.001385 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.001395 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.001412 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.001424 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.109009 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.109063 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.109074 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.109091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.109103 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.187741 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/2.log" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.188583 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/1.log" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.191229 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" exitCode=1 Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.191267 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.191350 4831 scope.go:117] "RemoveContainer" containerID="4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.191921 4831 scope.go:117] "RemoveContainer" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" Nov 24 08:16:15 crc kubenswrapper[4831]: E1124 08:16:15.192070 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.207647 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.211410 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.211440 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.211450 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.211465 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.211475 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.219331 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.229951 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.241572 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.254164 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.264701 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.279686 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.292680 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.304509 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.313809 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.313846 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.313855 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.313870 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.313879 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.316689 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.328018 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.338249 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.349049 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.366789 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.378031 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.387719 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.404495 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dddc328e7e5578374e66dfcd7448709ab94662410fa4b4ef49f47b577f0b91a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:15:57Z\\\",\\\"message\\\":\\\"t/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.885895 6152 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1124 08:15:57.886454 6152 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1124 08:15:57.886485 6152 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1124 08:15:57.886511 6152 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1124 08:15:57.886531 6152 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1124 08:15:57.886563 6152 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1124 08:15:57.886575 6152 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1124 08:15:57.886577 6152 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1124 08:15:57.886561 6152 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1124 08:15:57.886584 6152 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1124 08:15:57.886589 6152 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1124 08:15:57.886597 6152 handler.go:208] Removed *v1.Node event handler 2\\\\nI1124 08:15:57.886611 6152 factory.go:656] Stopping watch factory\\\\nI1124 08:15:57.886632 6152 ovnkube.go:599] Stopped ovnkube\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:15Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.415743 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.415777 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.415787 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.415803 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.415813 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.517308 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.517359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.517369 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.517384 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.517393 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.619266 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.619460 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.619477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.619501 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.619513 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.721764 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.721821 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.721832 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.721849 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.721859 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.823971 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.824018 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.824032 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.824050 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.824062 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.925777 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.925810 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.925825 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.925870 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:15 crc kubenswrapper[4831]: I1124 08:16:15.925881 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:15Z","lastTransitionTime":"2025-11-24T08:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.028427 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.028477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.028486 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.028500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.028509 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.130367 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.130405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.130417 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.130433 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.130444 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.195379 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/2.log" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.198625 4831 scope.go:117] "RemoveContainer" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.198851 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.211016 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.220904 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.229664 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.232101 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.232171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.232182 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.232200 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.232211 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.241571 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.267241 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.290510 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.309632 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.321723 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.333033 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.334449 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.334482 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.334493 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.334510 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.334521 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.344013 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.355603 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.364090 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.373820 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.391443 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.405571 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.414526 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.430148 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.436462 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.436498 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.436507 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.436521 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.436530 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.455984 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.456110 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.456183 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:32.456165283 +0000 UTC m=+66.331310426 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.538959 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.538997 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.539009 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.539024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.539035 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.626567 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.637180 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.641050 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.641079 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.641089 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.641104 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.641114 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.647524 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.658794 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.674619 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.684967 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.697333 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.711443 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.723211 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.735361 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.743144 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.743192 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.743203 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.743222 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.743233 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.748609 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.759606 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.772880 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.785547 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.804677 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.816009 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.825012 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.842017 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.845430 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.845480 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.845495 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.845513 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.845526 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.893008 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.893046 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.893067 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.893008 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.893131 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.893176 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.893265 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:16 crc kubenswrapper[4831]: E1124 08:16:16.893299 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.906073 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.918539 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.930369 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.944149 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.947610 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.947644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.947654 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.947669 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.947680 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:16Z","lastTransitionTime":"2025-11-24T08:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.955009 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.965615 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:16 crc kubenswrapper[4831]: I1124 08:16:16.983536 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.000138 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:16Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.011585 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.033512 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.044791 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.049556 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.049596 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.049606 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.049621 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.049633 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.056107 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.064529 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.075775 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.088764 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.099271 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.112850 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:17Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.152237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.152277 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.152288 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.152304 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.152334 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.254681 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.254730 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.254739 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.254756 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.254765 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.357559 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.357604 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.357615 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.357632 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.357643 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.459623 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.459690 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.459708 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.459741 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.459752 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.563174 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.563214 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.563222 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.563238 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.563249 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.665914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.665947 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.665958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.665983 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.665995 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.768453 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.768585 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.768596 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.768609 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.768618 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.871329 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.871365 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.871374 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.871387 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.871396 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.973665 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.973703 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.973712 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.973726 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:17 crc kubenswrapper[4831]: I1124 08:16:17.973736 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:17Z","lastTransitionTime":"2025-11-24T08:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.076952 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.077001 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.077010 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.077025 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.077035 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.179772 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.179833 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.179847 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.179868 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.179880 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.281890 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.281925 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.281933 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.281947 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.281956 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.384468 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.384523 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.384534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.384551 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.384563 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.486029 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.486084 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.486095 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.486112 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.486123 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.588408 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.588450 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.588458 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.588472 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.588481 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.673930 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674020 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:16:50.674001776 +0000 UTC m=+84.549146929 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.674063 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.674096 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.674128 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674197 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674245 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674256 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:50.674242993 +0000 UTC m=+84.549388166 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674261 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674277 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:50.674269584 +0000 UTC m=+84.549414737 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674281 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674296 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.674343 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:50.674333806 +0000 UTC m=+84.549478959 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.690698 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.690735 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.690745 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.690759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.690767 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.775833 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.776110 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.776146 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.776171 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.776269 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:16:50.776239949 +0000 UTC m=+84.651385142 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.792995 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.793038 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.793049 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.793066 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.793076 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.893069 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.893092 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.893069 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.893251 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.893182 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.893197 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.893395 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:18 crc kubenswrapper[4831]: E1124 08:16:18.893411 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.894552 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.894572 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.894581 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.894594 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.894604 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.996583 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.996644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.996659 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.996675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:18 crc kubenswrapper[4831]: I1124 08:16:18.996686 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:18Z","lastTransitionTime":"2025-11-24T08:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.099150 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.099193 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.099208 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.099229 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.099246 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.201251 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.201297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.201310 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.201353 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.201364 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.303106 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.303195 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.303207 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.303228 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.303242 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.405920 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.406157 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.406190 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.406221 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.406239 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.509788 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.509887 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.509898 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.509917 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.509927 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.612263 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.612330 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.612340 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.612354 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.612363 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.715002 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.715067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.715084 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.715110 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.715130 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.818261 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.818355 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.818373 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.818401 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.818420 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.920763 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.920819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.920830 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.920847 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:19 crc kubenswrapper[4831]: I1124 08:16:19.920860 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:19Z","lastTransitionTime":"2025-11-24T08:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.023301 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.023376 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.023388 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.023404 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.023413 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.125676 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.125904 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.126018 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.126098 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.126189 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.227885 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.227919 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.227931 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.227948 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.227958 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.330607 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.330641 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.330649 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.330662 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.330671 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.432829 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.432864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.432875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.432892 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.432907 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.535297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.535381 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.535399 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.535422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.535437 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.637608 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.637646 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.637657 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.637676 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.637686 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.740029 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.740080 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.740093 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.740110 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.740122 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.842729 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.842801 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.842815 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.842831 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.842843 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.892443 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.892531 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:20 crc kubenswrapper[4831]: E1124 08:16:20.892567 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:20 crc kubenswrapper[4831]: E1124 08:16:20.892636 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.892689 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:20 crc kubenswrapper[4831]: E1124 08:16:20.892734 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.892889 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:20 crc kubenswrapper[4831]: E1124 08:16:20.893010 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.945835 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.945873 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.945883 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.945898 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:20 crc kubenswrapper[4831]: I1124 08:16:20.945909 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:20Z","lastTransitionTime":"2025-11-24T08:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.052594 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.052653 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.052690 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.052797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.052840 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.155603 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.155644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.155657 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.155675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.155688 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.258890 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.258962 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.258982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.259011 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.259032 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.404364 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.404405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.404416 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.404433 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.404446 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.507814 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.507859 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.507869 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.507887 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.507899 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.610962 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.611016 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.611032 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.611057 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.611072 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.660630 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.672389 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.674289 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.686849 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.700595 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.712813 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.712858 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.712868 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.712882 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.712891 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.720198 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.732528 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.744935 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.758045 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.789010 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.810394 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.814418 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.814450 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.814458 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.814472 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.814482 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.826335 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.840685 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.851039 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.862480 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.875524 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.886040 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.899270 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.912737 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:21Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.916679 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.916723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.916732 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.916750 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:21 crc kubenswrapper[4831]: I1124 08:16:21.916764 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:21Z","lastTransitionTime":"2025-11-24T08:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.019569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.019642 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.019659 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.019685 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.019713 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.121742 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.121779 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.121789 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.121804 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.121816 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.224752 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.224791 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.224800 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.224817 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.224829 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.327039 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.327082 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.327091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.327112 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.327128 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.430243 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.430277 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.430285 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.430300 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.430309 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.532427 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.532454 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.532478 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.532491 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.532502 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.634293 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.634376 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.634390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.634405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.634415 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.736165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.736201 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.736210 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.736223 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.736231 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.838133 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.838166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.838175 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.838188 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.838197 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.892650 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.892689 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.892678 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.892830 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:22 crc kubenswrapper[4831]: E1124 08:16:22.892793 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:22 crc kubenswrapper[4831]: E1124 08:16:22.892938 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:22 crc kubenswrapper[4831]: E1124 08:16:22.892997 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:22 crc kubenswrapper[4831]: E1124 08:16:22.893058 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.940685 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.940727 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.940739 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.940757 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:22 crc kubenswrapper[4831]: I1124 08:16:22.940768 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:22Z","lastTransitionTime":"2025-11-24T08:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.042969 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.043012 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.043024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.043041 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.043052 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.146126 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.146181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.146189 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.146208 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.146220 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.170577 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.170626 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.170638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.170655 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.170667 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.182967 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:23Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.186303 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.186359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.186373 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.186389 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.186436 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.197240 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:23Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.200103 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.200131 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.200141 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.200155 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.200165 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.211345 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:23Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.214691 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.214723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.214735 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.214752 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.214762 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.227377 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:23Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.230398 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.230437 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.230446 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.230461 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.230470 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.240993 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:23Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:23 crc kubenswrapper[4831]: E1124 08:16:23.241170 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.248489 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.248537 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.248547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.248560 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.248571 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.350803 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.350854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.350866 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.350879 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.350888 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.453514 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.453547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.453557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.453572 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.453582 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.556301 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.556352 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.556362 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.556378 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.556388 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.658920 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.658977 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.658987 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.659002 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.659012 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.761299 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.761361 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.761372 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.761414 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.761428 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.863818 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.863893 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.863912 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.863943 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.863964 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.967569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.967613 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.967626 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.967646 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:23 crc kubenswrapper[4831]: I1124 08:16:23.967658 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:23Z","lastTransitionTime":"2025-11-24T08:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.070598 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.070626 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.070633 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.070645 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.070653 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.173917 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.173999 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.174021 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.174072 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.174110 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.278087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.278180 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.278204 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.278235 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.278256 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.381232 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.381273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.381281 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.381296 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.381305 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.485553 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.485611 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.485623 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.485644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.485658 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.587751 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.587788 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.587797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.587814 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.587841 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.689699 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.689727 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.689735 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.689747 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.689756 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.791644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.791675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.791688 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.791701 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.791709 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.892639 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.892676 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.892638 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:24 crc kubenswrapper[4831]: E1124 08:16:24.892810 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:24 crc kubenswrapper[4831]: E1124 08:16:24.892904 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:24 crc kubenswrapper[4831]: E1124 08:16:24.892977 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.893118 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:24 crc kubenswrapper[4831]: E1124 08:16:24.893183 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.894143 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.894173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.894181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.894193 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.894202 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.996142 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.996170 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.996178 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.996190 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:24 crc kubenswrapper[4831]: I1124 08:16:24.996198 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:24Z","lastTransitionTime":"2025-11-24T08:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.097759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.097787 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.097848 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.097864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.097873 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.199416 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.199501 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.199511 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.199524 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.199533 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.301249 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.301286 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.301294 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.301307 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.301333 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.403343 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.403373 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.403381 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.403393 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.403402 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.505500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.505546 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.505556 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.505568 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.505578 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.608500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.608526 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.608534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.608546 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.608554 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.712091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.712145 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.712156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.712172 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.712184 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.814746 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.814801 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.814812 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.814827 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.814839 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.916740 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.916783 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.916796 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.916811 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:25 crc kubenswrapper[4831]: I1124 08:16:25.916823 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:25Z","lastTransitionTime":"2025-11-24T08:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.019494 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.019542 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.019553 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.019573 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.019584 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.121694 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.121734 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.121744 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.121759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.121769 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.223636 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.223667 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.223675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.223689 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.223698 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.325583 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.325636 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.325647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.325664 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.325720 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.427831 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.427865 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.427876 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.427895 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.427905 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.529861 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.529911 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.529930 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.529954 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.529970 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.632451 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.632522 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.632536 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.632552 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.632596 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.734903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.734977 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.735014 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.735034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.735045 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.836984 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.837067 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.837088 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.837114 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.837132 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.892671 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.892679 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.892751 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.892770 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:26 crc kubenswrapper[4831]: E1124 08:16:26.893628 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:26 crc kubenswrapper[4831]: E1124 08:16:26.893808 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:26 crc kubenswrapper[4831]: E1124 08:16:26.894012 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:26 crc kubenswrapper[4831]: E1124 08:16:26.894296 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.910594 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.925367 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939008 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939245 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939272 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939288 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939310 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.939350 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:26Z","lastTransitionTime":"2025-11-24T08:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.950083 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.964488 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.975538 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:26 crc kubenswrapper[4831]: I1124 08:16:26.985439 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:26.999994 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:26Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.020048 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.031046 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.040959 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.041919 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.041961 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.041970 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.041985 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.042040 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.057179 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.069604 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.081961 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.093360 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.106114 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.119771 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.128336 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:27Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.144855 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.145172 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.145196 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.145216 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.145239 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.247489 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.247529 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.247545 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.247568 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.247585 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.350557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.350586 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.350594 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.350606 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.350616 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.452690 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.452715 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.452722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.452735 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.452744 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.555836 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.555880 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.555897 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.555914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.555926 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.658031 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.658096 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.658119 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.658181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.658196 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.760700 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.760784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.760823 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.760841 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.760854 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.863513 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.863537 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.863545 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.863557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.863566 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.966390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.966429 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.966438 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.966453 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:27 crc kubenswrapper[4831]: I1124 08:16:27.966464 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:27Z","lastTransitionTime":"2025-11-24T08:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.069124 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.069166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.069177 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.069199 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.069210 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.171026 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.171061 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.171069 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.171081 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.171090 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.272922 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.273197 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.273283 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.273390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.273464 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.375934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.375999 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.376013 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.376026 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.376036 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.478973 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.479006 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.479016 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.479029 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.479044 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.582180 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.582224 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.582233 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.582253 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.582265 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.685237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.685292 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.685305 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.685384 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.685407 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.788562 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.788604 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.788614 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.788633 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.788647 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.891721 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.891806 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.891819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.891943 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.891956 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.893058 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:28 crc kubenswrapper[4831]: E1124 08:16:28.893144 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.893268 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:28 crc kubenswrapper[4831]: E1124 08:16:28.893340 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.893465 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:28 crc kubenswrapper[4831]: E1124 08:16:28.893546 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.894375 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:28 crc kubenswrapper[4831]: E1124 08:16:28.894434 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.993662 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.993694 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.993703 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.993716 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:28 crc kubenswrapper[4831]: I1124 08:16:28.993725 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:28Z","lastTransitionTime":"2025-11-24T08:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.095708 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.095748 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.095759 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.095776 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.095788 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.198235 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.198279 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.198295 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.198382 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.198436 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.301834 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.301875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.301884 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.301900 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.301910 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.414724 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.414781 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.414804 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.414832 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.414853 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.517078 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.517129 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.517149 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.517173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.517190 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.623848 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.623879 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.623887 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.623899 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.623908 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.726404 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.726441 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.726449 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.726463 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.726472 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.828784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.828847 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.828860 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.828883 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.828896 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.931470 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.931509 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.931518 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.931532 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:29 crc kubenswrapper[4831]: I1124 08:16:29.931540 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:29Z","lastTransitionTime":"2025-11-24T08:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.033853 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.033908 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.033919 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.033935 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.033946 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.137168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.137218 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.137227 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.137245 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.137254 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.239919 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.239956 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.239965 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.239980 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.239990 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.342516 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.342625 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.342647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.342680 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.342701 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.445603 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.445643 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.445652 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.445666 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.445675 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.547985 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.548031 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.548040 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.548056 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.548066 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.650864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.650900 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.650909 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.650932 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.650942 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.753746 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.753779 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.753787 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.753800 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.753810 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.856005 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.856057 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.856068 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.856086 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.856097 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.895143 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.895345 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:30 crc kubenswrapper[4831]: E1124 08:16:30.895498 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.895548 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.895565 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:30 crc kubenswrapper[4831]: E1124 08:16:30.895838 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:30 crc kubenswrapper[4831]: E1124 08:16:30.895909 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.896117 4831 scope.go:117] "RemoveContainer" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" Nov 24 08:16:30 crc kubenswrapper[4831]: E1124 08:16:30.896145 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:30 crc kubenswrapper[4831]: E1124 08:16:30.896249 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.958370 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.958407 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.958415 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.958428 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:30 crc kubenswrapper[4831]: I1124 08:16:30.958437 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:30Z","lastTransitionTime":"2025-11-24T08:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.060359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.060394 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.060405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.060418 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.060426 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.162613 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.162668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.162679 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.162693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.162704 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.264856 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.264894 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.264902 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.264916 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.264924 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.367010 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.367046 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.367062 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.367079 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.367088 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.468788 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.468821 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.468829 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.468841 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.468852 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.571470 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.571524 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.571534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.571557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.571566 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.674181 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.674227 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.674237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.674254 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.674267 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.776511 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.776561 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.776572 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.776584 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.776592 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.879259 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.879297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.879309 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.879346 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.879358 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.981195 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.981233 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.981242 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.981258 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:31 crc kubenswrapper[4831]: I1124 08:16:31.981267 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:31Z","lastTransitionTime":"2025-11-24T08:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.083739 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.083770 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.083778 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.083790 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.083799 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.186305 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.186400 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.186408 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.186421 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.186431 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.288601 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.288639 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.288649 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.288664 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.288674 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.391289 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.391353 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.391366 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.391383 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.391394 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.493048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.493083 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.493093 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.493109 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.493121 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.507487 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.507646 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.507740 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:17:04.507719933 +0000 UTC m=+98.382865146 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.595758 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.595795 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.595803 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.595817 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.595826 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.697838 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.697878 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.697893 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.697909 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.697919 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.801025 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.801060 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.801069 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.801085 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.801097 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.892814 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.893202 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.892874 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.893507 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.893000 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.892841 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.893836 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:32 crc kubenswrapper[4831]: E1124 08:16:32.894088 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.903055 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.903083 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.903091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.903121 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:32 crc kubenswrapper[4831]: I1124 08:16:32.903130 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:32Z","lastTransitionTime":"2025-11-24T08:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.006272 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.006559 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.006644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.006717 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.006778 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.109825 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.110047 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.110165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.110259 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.110364 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.212364 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.212413 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.212422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.212436 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.212445 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.314622 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.314653 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.314662 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.314673 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.314682 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.408277 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.408331 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.408339 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.408353 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.408361 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.420542 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:33Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.423652 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.423756 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.423850 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.423939 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.424008 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.434820 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:33Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.438132 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.438156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.438165 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.438178 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.438203 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.450135 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:33Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.453437 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.453460 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.453469 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.453499 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.453508 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.464243 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:33Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.467351 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.467376 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.467385 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.467413 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.467421 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.484848 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:33Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:33 crc kubenswrapper[4831]: E1124 08:16:33.484960 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.486212 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.486251 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.486259 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.486286 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.486295 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.588012 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.588039 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.588049 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.588063 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.588073 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.691187 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.691223 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.691233 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.691268 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.691279 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.793455 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.793492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.793501 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.793515 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.793523 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.895953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.895986 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.895994 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.896021 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.896030 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.998169 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.998205 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.998213 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.998227 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:33 crc kubenswrapper[4831]: I1124 08:16:33.998237 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:33Z","lastTransitionTime":"2025-11-24T08:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.100578 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.100887 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.100985 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.101082 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.101165 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.204195 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.204256 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.204266 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.204278 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.204289 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.306436 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.306693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.306799 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.306898 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.306975 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.409221 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.409249 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.409259 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.409273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.409282 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.511748 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.511781 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.511789 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.511802 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.511810 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.613309 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.613357 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.613366 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.613382 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.613393 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.715534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.715578 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.715587 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.715603 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.715611 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.817772 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.817806 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.817816 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.817831 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.817841 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.892411 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:34 crc kubenswrapper[4831]: E1124 08:16:34.892738 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.892954 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:34 crc kubenswrapper[4831]: E1124 08:16:34.893072 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.893205 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.893402 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:34 crc kubenswrapper[4831]: E1124 08:16:34.893526 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:34 crc kubenswrapper[4831]: E1124 08:16:34.893412 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.920414 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.920481 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.920493 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.920512 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:34 crc kubenswrapper[4831]: I1124 08:16:34.920524 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:34Z","lastTransitionTime":"2025-11-24T08:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.024900 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.024953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.024967 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.024997 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.025012 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.128138 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.128188 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.128204 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.128223 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.128236 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.231536 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.231573 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.231581 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.231597 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.231609 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.262757 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/0.log" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.262817 4831 generic.go:334] "Generic (PLEG): container finished" podID="89aea50f-21a1-4854-a107-6e38c780166d" containerID="976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b" exitCode=1 Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.262842 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerDied","Data":"976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.263296 4831 scope.go:117] "RemoveContainer" containerID="976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.275348 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.287736 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.300705 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.313483 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"2025-11-24T08:15:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50\\\\n2025-11-24T08:15:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50 to /host/opt/cni/bin/\\\\n2025-11-24T08:15:50Z [verbose] multus-daemon started\\\\n2025-11-24T08:15:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:16:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.324098 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.334145 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.334177 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.334192 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.334206 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.334214 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.335168 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.349143 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.369622 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.381552 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.395053 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.415383 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.428769 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.436789 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.436813 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.436822 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.436836 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.436845 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.442677 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.453200 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.494031 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.523377 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.539448 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.539480 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.539493 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.539508 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.539519 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.551141 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.562899 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:35Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.641099 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.641265 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.641387 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.641501 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.641603 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.744231 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.744261 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.744270 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.744285 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.744296 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.846621 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.846658 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.846688 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.846708 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.846719 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.948456 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.948743 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.948827 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.948914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:35 crc kubenswrapper[4831]: I1124 08:16:35.948996 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:35Z","lastTransitionTime":"2025-11-24T08:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.050732 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.050807 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.050831 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.050927 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.050947 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.153862 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.154157 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.154234 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.154309 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.154415 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.257496 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.257547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.257557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.257576 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.257587 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.267281 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/0.log" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.267360 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerStarted","Data":"a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.282016 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.297648 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.310660 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.319136 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.332848 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.345441 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.357544 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.359712 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.359832 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.359975 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.360120 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.360283 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.367910 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.378790 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.389832 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.400731 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.413445 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"2025-11-24T08:15:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50\\\\n2025-11-24T08:15:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50 to /host/opt/cni/bin/\\\\n2025-11-24T08:15:50Z [verbose] multus-daemon started\\\\n2025-11-24T08:15:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:16:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.424100 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.441157 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.452850 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.462888 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.462928 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.462942 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.462960 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.462971 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.471859 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.482225 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.493910 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.565448 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.566397 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.566529 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.566619 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.566709 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.669479 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.669756 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.669827 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.669905 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.669986 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.772083 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.772338 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.772413 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.772481 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.772545 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.875430 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.875460 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.875471 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.875487 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.875498 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.896422 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:36 crc kubenswrapper[4831]: E1124 08:16:36.896578 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.896787 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:36 crc kubenswrapper[4831]: E1124 08:16:36.896876 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.897060 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:36 crc kubenswrapper[4831]: E1124 08:16:36.897149 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.897431 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:36 crc kubenswrapper[4831]: E1124 08:16:36.897530 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.910685 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.924152 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.940720 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.955935 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.968913 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.977953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.978001 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.978015 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.978033 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.978044 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:36Z","lastTransitionTime":"2025-11-24T08:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.982513 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"2025-11-24T08:15:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50\\\\n2025-11-24T08:15:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50 to /host/opt/cni/bin/\\\\n2025-11-24T08:15:50Z [verbose] multus-daemon started\\\\n2025-11-24T08:15:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:16:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:36 crc kubenswrapper[4831]: I1124 08:16:36.993529 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:36Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.013264 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.024256 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.042550 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.056565 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.074472 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.080509 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.080540 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.080551 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.080566 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.080577 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.088347 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.100878 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.117041 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.129218 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.142973 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.158997 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:37Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.183484 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.183832 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.183958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.184043 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.184114 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.287026 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.287070 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.287086 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.287105 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.287118 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.390606 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.390656 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.390678 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.390704 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.390722 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.492993 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.493073 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.493087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.493112 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.493123 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.597555 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.597990 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.598635 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.598765 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.598876 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.702875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.702925 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.702937 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.702953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.702963 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.805056 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.805103 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.805115 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.805131 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.805141 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.907654 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.907690 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.907716 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.907729 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:37 crc kubenswrapper[4831]: I1124 08:16:37.907757 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:37Z","lastTransitionTime":"2025-11-24T08:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.009828 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.009863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.009875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.009889 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.009899 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.111826 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.112127 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.112231 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.112339 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.112592 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.215653 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.215701 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.215710 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.215725 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.215736 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.318495 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.318544 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.318552 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.318571 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.318612 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.421765 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.421797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.421808 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.421833 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.421842 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.524159 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.524195 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.524205 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.524219 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.524228 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.626110 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.626149 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.626158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.626171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.626180 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.728134 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.728158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.728166 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.728179 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.728187 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.829876 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.829907 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.829915 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.829929 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.829939 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.892586 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.892641 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:38 crc kubenswrapper[4831]: E1124 08:16:38.892729 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.892756 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.892794 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:38 crc kubenswrapper[4831]: E1124 08:16:38.892945 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:38 crc kubenswrapper[4831]: E1124 08:16:38.893143 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:38 crc kubenswrapper[4831]: E1124 08:16:38.893211 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.939655 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.939714 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.939726 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.939747 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:38 crc kubenswrapper[4831]: I1124 08:16:38.939762 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:38Z","lastTransitionTime":"2025-11-24T08:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.042667 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.042722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.042733 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.042754 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.042766 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.144726 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.144770 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.144800 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.144816 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.144827 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.247264 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.247297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.247309 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.247339 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.247351 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.349764 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.349791 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.349799 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.349811 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.349820 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.452273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.452335 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.452348 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.452369 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.452380 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.553687 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.553723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.553736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.553753 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.553763 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.656039 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.656080 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.656093 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.656111 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.656121 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.758381 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.758411 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.758419 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.758432 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.758439 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.860370 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.860408 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.860416 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.860431 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.860440 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.962338 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.962374 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.962384 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.962399 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:39 crc kubenswrapper[4831]: I1124 08:16:39.962409 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:39Z","lastTransitionTime":"2025-11-24T08:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.065518 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.065558 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.065569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.065588 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.065606 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.167690 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.167736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.167749 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.167766 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.167778 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.269728 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.269761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.269771 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.269784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.269793 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.371814 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.371864 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.371875 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.371894 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.371907 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.473903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.473947 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.473960 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.473977 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.473988 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.576993 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.577024 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.577032 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.577045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.577063 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.679198 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.679260 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.679272 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.679312 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.679337 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.781492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.781532 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.781563 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.781581 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.781592 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.883362 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.883398 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.883409 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.883422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.883432 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.892800 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.892839 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:40 crc kubenswrapper[4831]: E1124 08:16:40.892899 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.892978 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.892801 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:40 crc kubenswrapper[4831]: E1124 08:16:40.893076 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:40 crc kubenswrapper[4831]: E1124 08:16:40.893127 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:40 crc kubenswrapper[4831]: E1124 08:16:40.893185 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.986045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.986091 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.986102 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.986118 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:40 crc kubenswrapper[4831]: I1124 08:16:40.986129 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:40Z","lastTransitionTime":"2025-11-24T08:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.088960 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.088992 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.089000 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.089013 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.089021 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.191573 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.191605 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.191613 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.191625 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.191634 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.293970 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.294007 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.294017 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.294032 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.294043 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.395668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.395693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.395701 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.395714 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.395724 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.497948 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.497984 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.497994 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.498009 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.498018 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.600151 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.600187 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.600198 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.600213 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.600226 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.702783 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.702810 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.702817 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.702830 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.702838 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.806123 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.806173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.806184 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.806201 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.806211 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.908807 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.908854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.908862 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.908876 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:41 crc kubenswrapper[4831]: I1124 08:16:41.908885 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:41Z","lastTransitionTime":"2025-11-24T08:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.011518 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.011551 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.011559 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.011589 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.011599 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.114297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.114348 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.114359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.114373 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.114383 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.217311 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.217378 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.217389 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.217405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.217417 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.319389 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.319423 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.319435 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.319452 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.319465 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.421570 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.421606 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.421617 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.421630 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.421667 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.523992 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.524039 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.524051 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.524070 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.524082 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.626678 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.626710 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.626719 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.626736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.626744 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.729229 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.729289 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.729307 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.729357 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.729378 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.832105 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.832158 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.832177 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.832202 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.832220 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.892883 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.892938 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.893015 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.893063 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:42 crc kubenswrapper[4831]: E1124 08:16:42.893064 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:42 crc kubenswrapper[4831]: E1124 08:16:42.893197 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:42 crc kubenswrapper[4831]: E1124 08:16:42.893272 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:42 crc kubenswrapper[4831]: E1124 08:16:42.893368 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.934016 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.934088 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.934101 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.934114 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:42 crc kubenswrapper[4831]: I1124 08:16:42.934125 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:42Z","lastTransitionTime":"2025-11-24T08:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.037034 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.037088 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.037097 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.037113 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.037122 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.139761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.139798 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.139834 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.139851 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.139861 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.241987 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.242031 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.242047 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.242069 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.242082 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.345565 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.345601 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.345611 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.345625 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.345701 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.447719 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.447854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.447994 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.448021 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.448036 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.550171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.550214 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.550223 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.550237 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.550248 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.557314 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.557377 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.557387 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.557402 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.557413 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.570750 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.574497 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.574556 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.574569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.574584 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.574593 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.586520 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.590125 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.590163 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.590173 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.590192 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.590207 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.605913 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.611762 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.611996 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.612103 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.612170 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.612200 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.632515 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.636889 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.637077 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.637172 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.637272 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.637411 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.651157 4831 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"59b72184-9079-421a-8fa8-bad568411b37\\\",\\\"systemUUID\\\":\\\"f42275a4-fa6b-4c08-99eb-479953ecb2b1\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:43Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:43 crc kubenswrapper[4831]: E1124 08:16:43.651681 4831 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.653196 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.653379 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.653492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.653585 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.653683 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.756161 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.756222 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.756239 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.756261 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.756279 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.859153 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.859183 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.859193 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.859205 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.859213 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.961093 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.961132 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.961143 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.961159 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:43 crc kubenswrapper[4831]: I1124 08:16:43.961169 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:43Z","lastTransitionTime":"2025-11-24T08:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.063834 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.063909 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.063933 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.063961 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.063990 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.167390 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.167424 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.167433 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.167448 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.167460 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.269608 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.269647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.269658 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.269675 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.269687 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.372138 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.372179 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.372190 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.372207 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.372219 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.474409 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.474443 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.474454 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.474477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.474486 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.576416 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.576458 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.576466 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.576481 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.576490 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.678129 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.678628 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.678698 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.678769 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.678840 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.781102 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.781148 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.781159 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.781196 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.781210 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.883693 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.883728 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.883736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.883749 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.883764 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.892933 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.893052 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.893152 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:44 crc kubenswrapper[4831]: E1124 08:16:44.893153 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:44 crc kubenswrapper[4831]: E1124 08:16:44.893253 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:44 crc kubenswrapper[4831]: E1124 08:16:44.893374 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.893084 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:44 crc kubenswrapper[4831]: E1124 08:16:44.893498 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.986294 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.986357 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.986367 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.986405 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:44 crc kubenswrapper[4831]: I1124 08:16:44.986418 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:44Z","lastTransitionTime":"2025-11-24T08:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.088757 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.088792 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.088800 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.088815 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.088823 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.190885 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.191097 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.191187 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.191278 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.191399 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.293232 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.293288 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.293297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.293311 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.293335 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.395182 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.395508 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.395585 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.395655 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.395717 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.497901 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.497927 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.497934 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.497948 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.497958 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.599949 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.600019 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.600033 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.600052 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.600067 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.701995 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.702036 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.702045 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.702060 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.702070 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.804399 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.804457 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.804469 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.804483 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.804494 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.893513 4831 scope.go:117] "RemoveContainer" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.905973 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.906033 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.906042 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.906056 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:45 crc kubenswrapper[4831]: I1124 08:16:45.906065 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:45Z","lastTransitionTime":"2025-11-24T08:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.008113 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.008353 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.008448 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.008534 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.008615 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.111116 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.111340 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.111477 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.111563 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.111625 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.213201 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.213243 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.213254 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.213269 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.213285 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.315463 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.315749 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.315857 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.315949 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.316044 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.423141 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.423168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.423179 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.423194 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.423204 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.524932 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.524958 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.524968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.524982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.524990 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.627460 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.627499 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.627509 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.627522 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.627531 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.730535 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.730571 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.730578 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.730591 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.730600 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.833127 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.833164 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.833172 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.833187 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.833196 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.892897 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.892912 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:46 crc kubenswrapper[4831]: E1124 08:16:46.893283 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.892966 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.892924 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:46 crc kubenswrapper[4831]: E1124 08:16:46.893433 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:46 crc kubenswrapper[4831]: E1124 08:16:46.893497 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:46 crc kubenswrapper[4831]: E1124 08:16:46.893539 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.905683 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.918199 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.929534 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.935638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.935672 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.935700 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.935716 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.935727 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:46Z","lastTransitionTime":"2025-11-24T08:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.946825 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"2025-11-24T08:15:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50\\\\n2025-11-24T08:15:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50 to /host/opt/cni/bin/\\\\n2025-11-24T08:15:50Z [verbose] multus-daemon started\\\\n2025-11-24T08:15:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:16:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.956730 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.967288 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:46 crc kubenswrapper[4831]: I1124 08:16:46.979163 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:46Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.005055 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.015528 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.026602 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.038080 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.038118 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.038129 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.038142 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.038153 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.044368 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.057514 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.069285 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.078529 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.091221 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.104507 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.114462 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.127843 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.140547 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.140569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.140577 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.140589 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.140599 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.243669 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.244419 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.244464 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.244480 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.244490 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.300032 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/3.log" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.300725 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/2.log" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.303732 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" exitCode=1 Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.303785 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.303828 4831 scope.go:117] "RemoveContainer" containerID="2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.304785 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:16:47 crc kubenswrapper[4831]: E1124 08:16:47.305002 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.317853 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7aa4c55e44b6a88fbb2293356408fd19d26c9910933facd2814e25e0a1515b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.330490 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.342534 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-k6hd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"89aea50f-21a1-4854-a107-6e38c780166d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:35Z\\\",\\\"message\\\":\\\"2025-11-24T08:15:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50\\\\n2025-11-24T08:15:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_f10623d2-bf5e-435a-bd06-912b3f3a1b50 to /host/opt/cni/bin/\\\\n2025-11-24T08:15:50Z [verbose] multus-daemon started\\\\n2025-11-24T08:15:50Z [verbose] Readiness Indicator file check\\\\n2025-11-24T08:16:35Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bj7wc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-k6hd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.346135 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.346162 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.346171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.346183 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.346192 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.351762 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-mnrjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9da0997c-8ea4-4af7-b1db-428d24f7d635\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a4f04d6a587614bf1c0b4902cfca5dc8f270b4bf33ac7ef146d7a1ea2fe107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khwh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:49Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-mnrjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.361992 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa82b248-7cf0-4285-8978-9ef7d10a265b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b008735e3dd730c76b4faee93ca04935bb7ea4c2681018cc5c4cb1a8ae23065\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c10e85d93df874200b7ac2d01adb705a13b16e258e2d957a588db647da2e1cdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9bjbt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:59Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2dnft\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.374680 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c9ffde08-9f06-4131-be02-130e559374de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91124a4d363a3d2177222f6623c13c7f39814d94c6591f4f10fcabb70e47b845\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c8bbbcc4191c49ffcc6702bb0a4cabd31c0b8021189af15409cf3cb88161dba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e4d58c7db00754029498ff691b7afb1877a3209785a5e36679d0bd44b2480a53\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cdcf27778285a495fcf615ad1ae08f7b3e5bcca1f67a416893f54d62c4b73bbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.394290 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2d2ec6a7-012a-46d6-bcaf-0ec7b01608eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85b554efa305ed9a81015c176d2b6f1d3838d30bbfadc817e89b212537e8eaec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f63d885faedcffd9a60a212e9dd3b6783f4cfe5d944f246994349ae8109ec16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8460209fe01008c585c31ed13433ea1e6f530169939c0a443cdefd15b39a27bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1b675bfcc2025e4c2447342492a015b0b4ea495a1587af2542bdbc326b90473\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b075ee24c597c7f4c2b9fc278d2e1365bb440e1d5d65256dfd622f08b02c215a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fc8625edeaba857a47e2b181270dee4b20fc076913e8f1e85f2a60ad450cd89\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ffedefb5b6b7bf7acaa1306b492265bb4d079aebd9d1a1e761b7216caef014e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9168df03bdc3418b38bc89f2bf0722eb5de5652688c82d46d6accea208bc6574\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.405446 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.416804 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b857b1a-bb8a-4608-8643-e0a94ed82a1b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://17c85d6252ff04e9ff613493c249b49a22df36ecb43e7e82d0bcab2be0065028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7ztt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ftxxx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.434359 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77826b29-ac89-4b2a-afd9-746c0b1c13ff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2371f66ecb0af91910eb4e7ac1cc71961723b2833a1e83bbd989bf440b944c97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:14Z\\\",\\\"message\\\":\\\"ss event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1124 08:16:14.730675 6383 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:14.730686 6383 lb_config.go:1031] Cluster endpoints for openshift-kube-storage-version-migrator-operator/metrics for network=default are: map[]\\\\nI1124 08:16:14.730700 6383 services_controller.go:443] Built service openshift-kube-storage-version-migrator-operator/metrics LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.36\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1124 08:16:14.730721 6383 services_controller.go:444] Built service openshift-kube-storage-version-migrator-operator/metrics LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1124 08:16:14.730728 6383 services_controller.go:445] Built service openshift-kube-storage-version-migrator-operator/metrics LB template configs for network=default: []services.lbConfig(nil)\\\\nF1124 08:16:14.730737 6383 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-24T08:16:47Z\\\",\\\"message\\\":\\\"urable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1124 08:16:47.226107 6748 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-hxqjw\\\\nI1124 08:16:47.226921 6748 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-hxqjw\\\\nI1124 08:16:47.226927 6748 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-hxqjw in node crc\\\\nI1124 08:16:47.226931 6748 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-hxqjw after 0 failed attempt(s)\\\\nI1124 08:16:47.226935 6748 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-hxqjw\\\\nI1124 08:16:47.226947 6748 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1124 08:16:47.226956 6748 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1124 08:16:47.227001 6748 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:16:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h8gjg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:47Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-hxqjw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.444576 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52db3125-fb2e-4aa0-92de-7d5ef83c1ebd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aa20dc4c206339c061f4f9410e715acd95dd4919e497d32d06c9a93206f9ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0df51153f5b665f0738818c5637a2cc21ccc41f5404cd459a7a38cc6e7bc036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c619233e3465118916fd9d2e8e301ac224f658969a69c0b3842cc444b1e18f3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b67aa94aea3f6e577d77ef93f13b27126902cb080f7082b64887ec1df0827a50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.452141 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.452171 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.452179 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.452191 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.452200 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.455789 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://98258840580b9fffca661fafca91a485ad58471c10082131d740af7a2da0270b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.466700 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-556kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8de0f5f-48af-4be5-a67e-577d6f2897f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://39df6c2f4294a3ed9fb9b6e6202a1dfe6c8ccbbc5d118e63effd45123dcb49e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ht772\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-556kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.477057 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:46Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.489031 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"66dfbe4f-ed46-4f54-b304-427a18c7da39\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f51605c93fb6c83b015e4cd710067c32a507a137f3b8c024dbda0049c561b71d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3169b880b4978d20a15dba2f79cdd26b5b8bb516f21cc4ffa8e9de961b22fe23\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4342e168879c72d3e062c8144882ff011a09f67f1be4322f8555ef4da772c3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e065d4ea5bc0dd0ad612d9d534c4a533db4220ef52e2d7e8994e46e9adfce6a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb1fc2258ca37fb562c0b9753c55bd28721eecf76058ddff503b963dd2155faa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://404bbc57cd32d1f34581ad9793f82c4093391de9c6636691261ceb5a61ce965e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82a2336b868fddb464f3a689ad31d140d7839444932e9f3ea50dffd7533c7404\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jkfp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:46Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-kdjrf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.497647 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-k45nx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"650855b7-20e6-4261-a212-4bc75a671b96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mh6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:16:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-k45nx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.506887 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2cc7dc593ba061ee9af320906b7fbac2aece9498e61f82fe443f3fc6b7442ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1114601042bf76addbe6d4b3a34d8ebb525b40de1c1cd973188d7459990977d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.518242 4831 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5550ba22-2716-45bd-8dce-b90a44081892\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:16:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-24T08:15:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27eed08bbdabaa0c268b6e2e86a1bd06724b5bb0d37dce61119b8528f1b67e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3d9d3ebac459af8645b996c0eea2e17103a84480e92fccad9f762c5392dc4d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f24126d7481dd14e353ce6676d5885a54748cfb2b6e03c5a5994b5476f0ab04\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2fcfbbaaddf665932b8687b60ee9786455ddbde87dda98917d23f53ec9d01d25\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://589dce0c234cad9e2160c65863d0be827e98cb359b7d39967bd48ee1c92fc117\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-24T08:15:47Z\\\",\\\"message\\\":\\\"le observer\\\\nW1124 08:15:46.578209 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1124 08:15:46.578380 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1124 08:15:46.579147 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2354164174/tls.crt::/tmp/serving-cert-2354164174/tls.key\\\\\\\"\\\\nI1124 08:15:47.557820 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1124 08:15:47.561726 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1124 08:15:47.561750 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1124 08:15:47.561768 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1124 08:15:47.561773 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1124 08:15:47.568438 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1124 08:15:47.568465 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568470 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1124 08:15:47.568473 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1124 08:15:47.568476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1124 08:15:47.568478 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1124 08:15:47.568480 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1124 08:15:47.568578 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1124 08:15:47.570878 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:16:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b9309186a2923abd78f0dc4093bcef45ba89b2ab9bb06ad466c83542a9c56a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-24T08:15:29Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccbc1e0040c548ca34b8e22aa25837bffb95c4364b262cabcb731a8459a8de63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-24T08:15:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-24T08:15:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-24T08:15:26Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-24T08:16:47Z is after 2025-08-24T17:21:41Z" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.553828 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.553877 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.553888 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.553905 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.553915 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.655478 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.655505 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.655514 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.655526 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.655534 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.758055 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.758090 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.758101 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.758117 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.758128 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.860236 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.860273 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.860285 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.860302 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.860313 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.962328 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.962363 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.962377 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.962393 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:47 crc kubenswrapper[4831]: I1124 08:16:47.962403 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:47Z","lastTransitionTime":"2025-11-24T08:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.064775 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.064807 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.064814 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.064829 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.064839 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.167120 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.167157 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.167168 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.167184 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.167196 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.268939 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.268973 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.268986 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.269001 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.269011 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.307213 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/3.log" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.370500 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.370529 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.370540 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.370555 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.370566 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.472525 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.472575 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.472583 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.472594 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.472603 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.578797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.578858 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.578885 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.578900 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.578910 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.680532 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.680564 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.680579 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.680596 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.680606 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.782924 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.782960 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.782968 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.782982 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.782992 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.885468 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.885504 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.885514 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.885528 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.885538 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.893442 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.893507 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:48 crc kubenswrapper[4831]: E1124 08:16:48.893550 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:48 crc kubenswrapper[4831]: E1124 08:16:48.893614 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.893676 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:48 crc kubenswrapper[4831]: E1124 08:16:48.893730 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.893794 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:48 crc kubenswrapper[4831]: E1124 08:16:48.893851 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.987550 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.987584 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.987644 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.987657 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:48 crc kubenswrapper[4831]: I1124 08:16:48.987665 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:48Z","lastTransitionTime":"2025-11-24T08:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.089933 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.089970 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.089988 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.090004 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.090016 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.192307 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.192351 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.192359 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.192373 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.192382 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.294784 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.294819 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.294827 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.294839 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.294848 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.396769 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.396810 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.396820 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.396835 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.396846 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.498649 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.498698 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.498708 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.498723 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.498732 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.600354 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.600388 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.600397 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.600409 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.600421 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.702801 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.702850 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.702861 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.702873 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.702881 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.804761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.804805 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.804816 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.804833 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.804844 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.907256 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.907297 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.907308 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.907341 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:49 crc kubenswrapper[4831]: I1124 08:16:49.907353 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:49Z","lastTransitionTime":"2025-11-24T08:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.009732 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.009770 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.009781 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.009797 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.009808 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.112663 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.112697 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.112708 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.112722 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.112737 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.214769 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.214833 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.214842 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.214856 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.214866 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.316447 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.316492 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.316504 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.316520 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.316536 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.418647 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.418689 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.418701 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.418715 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.418724 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.520670 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.520703 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.520710 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.520724 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.520753 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.623061 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.623100 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.623109 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.623123 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.623132 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.691045 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.691154 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.691201 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.691224 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691292 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.691278644 +0000 UTC m=+148.566423797 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691358 4831 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691388 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691386 4831 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691423 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.691410658 +0000 UTC m=+148.566555811 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691476 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.691454479 +0000 UTC m=+148.566599712 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691401 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691507 4831 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.691551 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.691539872 +0000 UTC m=+148.566685105 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.725761 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.725795 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.725805 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.725822 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.725832 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.792241 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.792444 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.792470 4831 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.792482 4831 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.792533 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.792519985 +0000 UTC m=+148.667665148 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.828033 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.828075 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.828083 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.828098 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.828110 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.892957 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.893081 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.893280 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.893436 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.893536 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.893588 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.893721 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:50 crc kubenswrapper[4831]: E1124 08:16:50.893853 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.933146 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.933533 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.933549 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.933619 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:50 crc kubenswrapper[4831]: I1124 08:16:50.933637 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:50Z","lastTransitionTime":"2025-11-24T08:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.036577 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.036629 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.036638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.036649 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.036658 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.138914 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.138953 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.138963 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.138979 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.138988 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.241194 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.241254 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.241265 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.241302 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.241330 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.343183 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.343210 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.343218 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.343230 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.343238 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.445806 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.445839 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.445849 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.445865 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.445876 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.548108 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.548147 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.548156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.548172 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.548181 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.651037 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.651075 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.651085 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.651100 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.651111 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.753798 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.753853 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.753863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.753878 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.753901 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.856679 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.856725 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.856736 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.856752 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.856763 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.958820 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.958854 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.958867 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.958886 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:51 crc kubenswrapper[4831]: I1124 08:16:51.958897 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:51Z","lastTransitionTime":"2025-11-24T08:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.061000 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.061029 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.061037 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.061048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.061057 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.163263 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.163296 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.163304 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.163342 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.163352 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.265711 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.265754 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.265762 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.265777 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.265785 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.368263 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.368298 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.368310 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.368343 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.368355 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.470601 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.470641 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.470653 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.470668 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.470679 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.574005 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.574046 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.574058 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.574076 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.574087 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.676544 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.676607 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.676627 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.676648 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.676664 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.778805 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.778838 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.778849 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.778863 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.778875 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.881285 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.881311 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.881348 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.881360 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.881370 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.892550 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.892581 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.892589 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.892556 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:52 crc kubenswrapper[4831]: E1124 08:16:52.892664 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:52 crc kubenswrapper[4831]: E1124 08:16:52.892770 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:52 crc kubenswrapper[4831]: E1124 08:16:52.893163 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:52 crc kubenswrapper[4831]: E1124 08:16:52.893449 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.983422 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.983502 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.983529 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.983557 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:52 crc kubenswrapper[4831]: I1124 08:16:52.983578 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:52Z","lastTransitionTime":"2025-11-24T08:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.085574 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.085616 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.085623 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.085638 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.085647 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.188345 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.188377 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.188389 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.188404 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.188439 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.290569 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.290613 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.290625 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.290641 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.290655 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.393911 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.394006 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.394027 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.394048 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.394064 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.496156 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.496186 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.496194 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.496206 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.496215 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.597874 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.597903 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.597920 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.597937 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.597946 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.700137 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.700192 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.700207 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.700225 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.700240 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.739053 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.739087 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.739098 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.739114 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.739125 4831 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-24T08:16:53Z","lastTransitionTime":"2025-11-24T08:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.791227 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk"] Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.791741 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.793658 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.793896 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.794218 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.795623 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.822675 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=64.822655231 podStartE2EDuration="1m4.822655231s" podCreationTimestamp="2025-11-24 08:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:53.82087549 +0000 UTC m=+87.696020653" watchObservedRunningTime="2025-11-24 08:16:53.822655231 +0000 UTC m=+87.697800384" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.877828 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podStartSLOduration=67.877748368 podStartE2EDuration="1m7.877748368s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:53.852966763 +0000 UTC m=+87.728111936" watchObservedRunningTime="2025-11-24 08:16:53.877748368 +0000 UTC m=+87.752893561" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.891964 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=32.891937102 podStartE2EDuration="32.891937102s" podCreationTimestamp="2025-11-24 08:16:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:53.891604853 +0000 UTC m=+87.766750006" watchObservedRunningTime="2025-11-24 08:16:53.891937102 +0000 UTC m=+87.767082295" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.924169 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b61dd28-b8ca-4ea5-8859-8e9517747564-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.924500 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.924587 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b61dd28-b8ca-4ea5-8859-8e9517747564-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.924660 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b61dd28-b8ca-4ea5-8859-8e9517747564-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.924736 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.925084 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-556kb" podStartSLOduration=67.925050794 podStartE2EDuration="1m7.925050794s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:53.924860859 +0000 UTC m=+87.800006092" watchObservedRunningTime="2025-11-24 08:16:53.925050794 +0000 UTC m=+87.800195997" Nov 24 08:16:53 crc kubenswrapper[4831]: I1124 08:16:53.993252 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-kdjrf" podStartSLOduration=67.993230063 podStartE2EDuration="1m7.993230063s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:53.993034117 +0000 UTC m=+87.868179280" watchObservedRunningTime="2025-11-24 08:16:53.993230063 +0000 UTC m=+87.868375216" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.025964 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b61dd28-b8ca-4ea5-8859-8e9517747564-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.026047 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.026070 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b61dd28-b8ca-4ea5-8859-8e9517747564-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.026091 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b61dd28-b8ca-4ea5-8859-8e9517747564-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.026126 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.026204 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.027005 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3b61dd28-b8ca-4ea5-8859-8e9517747564-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.027717 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b61dd28-b8ca-4ea5-8859-8e9517747564-service-ca\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.033186 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b61dd28-b8ca-4ea5-8859-8e9517747564-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.042244 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3b61dd28-b8ca-4ea5-8859-8e9517747564-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-j6vfk\" (UID: \"3b61dd28-b8ca-4ea5-8859-8e9517747564\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.052157 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=66.05213849 podStartE2EDuration="1m6.05213849s" podCreationTimestamp="2025-11-24 08:15:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.039213392 +0000 UTC m=+87.914358565" watchObservedRunningTime="2025-11-24 08:16:54.05213849 +0000 UTC m=+87.927283643" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.079467 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-k6hd5" podStartSLOduration=68.079452107 podStartE2EDuration="1m8.079452107s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.079108607 +0000 UTC m=+87.954253760" watchObservedRunningTime="2025-11-24 08:16:54.079452107 +0000 UTC m=+87.954597260" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.089632 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-mnrjr" podStartSLOduration=68.089615146 podStartE2EDuration="1m8.089615146s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.089128232 +0000 UTC m=+87.964273385" watchObservedRunningTime="2025-11-24 08:16:54.089615146 +0000 UTC m=+87.964760289" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.113157 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.127229 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2dnft" podStartSLOduration=68.127204765 podStartE2EDuration="1m8.127204765s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.103926623 +0000 UTC m=+87.979071776" watchObservedRunningTime="2025-11-24 08:16:54.127204765 +0000 UTC m=+88.002349918" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.325865 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" event={"ID":"3b61dd28-b8ca-4ea5-8859-8e9517747564","Type":"ContainerStarted","Data":"5198833fe0f3d5621efc3f4531f3a9b32c1285404a93c40ebe4c82564795e18f"} Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.325905 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" event={"ID":"3b61dd28-b8ca-4ea5-8859-8e9517747564","Type":"ContainerStarted","Data":"ccae05788c9ce079a78b77728881abde35a5f070b3375a80a082e9180877888c"} Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.338140 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.338120576 podStartE2EDuration="1m5.338120576s" podCreationTimestamp="2025-11-24 08:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.147923645 +0000 UTC m=+88.023068808" watchObservedRunningTime="2025-11-24 08:16:54.338120576 +0000 UTC m=+88.213265729" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.338654 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-j6vfk" podStartSLOduration=68.338648511 podStartE2EDuration="1m8.338648511s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:16:54.337449007 +0000 UTC m=+88.212594180" watchObservedRunningTime="2025-11-24 08:16:54.338648511 +0000 UTC m=+88.213793664" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.892785 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:54 crc kubenswrapper[4831]: E1124 08:16:54.892926 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.893091 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:54 crc kubenswrapper[4831]: E1124 08:16:54.893137 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.893570 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:54 crc kubenswrapper[4831]: I1124 08:16:54.893765 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:54 crc kubenswrapper[4831]: E1124 08:16:54.893937 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:54 crc kubenswrapper[4831]: E1124 08:16:54.894488 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.430709 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.431450 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:16:56 crc kubenswrapper[4831]: E1124 08:16:56.431616 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.892524 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.892537 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.892545 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:56 crc kubenswrapper[4831]: I1124 08:16:56.892579 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:56 crc kubenswrapper[4831]: E1124 08:16:56.893419 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:56 crc kubenswrapper[4831]: E1124 08:16:56.893520 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:56 crc kubenswrapper[4831]: E1124 08:16:56.893590 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:56 crc kubenswrapper[4831]: E1124 08:16:56.893662 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:58 crc kubenswrapper[4831]: I1124 08:16:58.893189 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:16:58 crc kubenswrapper[4831]: I1124 08:16:58.893189 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:16:58 crc kubenswrapper[4831]: I1124 08:16:58.893184 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:16:58 crc kubenswrapper[4831]: I1124 08:16:58.893222 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:16:58 crc kubenswrapper[4831]: E1124 08:16:58.894604 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:16:58 crc kubenswrapper[4831]: E1124 08:16:58.894814 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:16:58 crc kubenswrapper[4831]: E1124 08:16:58.894951 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:16:58 crc kubenswrapper[4831]: E1124 08:16:58.895106 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:16:58 crc kubenswrapper[4831]: I1124 08:16:58.911051 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 24 08:17:00 crc kubenswrapper[4831]: I1124 08:17:00.893134 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:00 crc kubenswrapper[4831]: I1124 08:17:00.893157 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:00 crc kubenswrapper[4831]: E1124 08:17:00.893736 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:00 crc kubenswrapper[4831]: I1124 08:17:00.893256 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:00 crc kubenswrapper[4831]: E1124 08:17:00.893852 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:00 crc kubenswrapper[4831]: I1124 08:17:00.893200 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:00 crc kubenswrapper[4831]: E1124 08:17:00.893995 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:00 crc kubenswrapper[4831]: E1124 08:17:00.894073 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:02 crc kubenswrapper[4831]: I1124 08:17:02.892907 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:02 crc kubenswrapper[4831]: I1124 08:17:02.893011 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:02 crc kubenswrapper[4831]: E1124 08:17:02.893030 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:02 crc kubenswrapper[4831]: I1124 08:17:02.892907 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:02 crc kubenswrapper[4831]: E1124 08:17:02.893176 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:02 crc kubenswrapper[4831]: E1124 08:17:02.893229 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:02 crc kubenswrapper[4831]: I1124 08:17:02.892908 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:02 crc kubenswrapper[4831]: E1124 08:17:02.893287 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:04 crc kubenswrapper[4831]: I1124 08:17:04.561125 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.561297 4831 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.561782 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs podName:650855b7-20e6-4261-a212-4bc75a671b96 nodeName:}" failed. No retries permitted until 2025-11-24 08:18:08.561763873 +0000 UTC m=+162.436909026 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs") pod "network-metrics-daemon-k45nx" (UID: "650855b7-20e6-4261-a212-4bc75a671b96") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 24 08:17:04 crc kubenswrapper[4831]: I1124 08:17:04.893272 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:04 crc kubenswrapper[4831]: I1124 08:17:04.893365 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:04 crc kubenswrapper[4831]: I1124 08:17:04.893403 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.893443 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.893543 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.893603 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:04 crc kubenswrapper[4831]: I1124 08:17:04.893828 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:04 crc kubenswrapper[4831]: E1124 08:17:04.894034 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:06 crc kubenswrapper[4831]: I1124 08:17:06.892851 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:06 crc kubenswrapper[4831]: I1124 08:17:06.892904 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:06 crc kubenswrapper[4831]: I1124 08:17:06.892923 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:06 crc kubenswrapper[4831]: I1124 08:17:06.892962 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:06 crc kubenswrapper[4831]: E1124 08:17:06.894085 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:06 crc kubenswrapper[4831]: E1124 08:17:06.894729 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:06 crc kubenswrapper[4831]: E1124 08:17:06.894882 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:06 crc kubenswrapper[4831]: E1124 08:17:06.894988 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:06 crc kubenswrapper[4831]: I1124 08:17:06.905208 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=8.905193215 podStartE2EDuration="8.905193215s" podCreationTimestamp="2025-11-24 08:16:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:06.904065683 +0000 UTC m=+100.779210846" watchObservedRunningTime="2025-11-24 08:17:06.905193215 +0000 UTC m=+100.780338358" Nov 24 08:17:08 crc kubenswrapper[4831]: I1124 08:17:08.893453 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:08 crc kubenswrapper[4831]: E1124 08:17:08.893621 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:08 crc kubenswrapper[4831]: I1124 08:17:08.893660 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:08 crc kubenswrapper[4831]: I1124 08:17:08.893752 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:08 crc kubenswrapper[4831]: E1124 08:17:08.893927 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:08 crc kubenswrapper[4831]: I1124 08:17:08.893983 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:08 crc kubenswrapper[4831]: E1124 08:17:08.894112 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:08 crc kubenswrapper[4831]: E1124 08:17:08.894199 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:10 crc kubenswrapper[4831]: I1124 08:17:10.893567 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:10 crc kubenswrapper[4831]: E1124 08:17:10.893714 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:10 crc kubenswrapper[4831]: I1124 08:17:10.894034 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:10 crc kubenswrapper[4831]: E1124 08:17:10.894094 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:10 crc kubenswrapper[4831]: I1124 08:17:10.894802 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:17:10 crc kubenswrapper[4831]: E1124 08:17:10.894942 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:17:10 crc kubenswrapper[4831]: I1124 08:17:10.895059 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:10 crc kubenswrapper[4831]: E1124 08:17:10.895117 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:10 crc kubenswrapper[4831]: I1124 08:17:10.895249 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:10 crc kubenswrapper[4831]: E1124 08:17:10.895353 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:12 crc kubenswrapper[4831]: I1124 08:17:12.893423 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:12 crc kubenswrapper[4831]: I1124 08:17:12.893481 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:12 crc kubenswrapper[4831]: I1124 08:17:12.893493 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:12 crc kubenswrapper[4831]: I1124 08:17:12.893530 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:12 crc kubenswrapper[4831]: E1124 08:17:12.893576 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:12 crc kubenswrapper[4831]: E1124 08:17:12.893818 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:12 crc kubenswrapper[4831]: E1124 08:17:12.893882 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:12 crc kubenswrapper[4831]: E1124 08:17:12.893981 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:14 crc kubenswrapper[4831]: I1124 08:17:14.892481 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:14 crc kubenswrapper[4831]: E1124 08:17:14.892653 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:14 crc kubenswrapper[4831]: I1124 08:17:14.892967 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:14 crc kubenswrapper[4831]: E1124 08:17:14.893061 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:14 crc kubenswrapper[4831]: I1124 08:17:14.893479 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:14 crc kubenswrapper[4831]: I1124 08:17:14.893567 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:14 crc kubenswrapper[4831]: E1124 08:17:14.893676 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:14 crc kubenswrapper[4831]: E1124 08:17:14.893886 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:16 crc kubenswrapper[4831]: I1124 08:17:16.892762 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:16 crc kubenswrapper[4831]: I1124 08:17:16.892821 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:16 crc kubenswrapper[4831]: I1124 08:17:16.892856 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:16 crc kubenswrapper[4831]: I1124 08:17:16.892821 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:16 crc kubenswrapper[4831]: E1124 08:17:16.893947 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:16 crc kubenswrapper[4831]: E1124 08:17:16.894006 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:16 crc kubenswrapper[4831]: E1124 08:17:16.894073 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:16 crc kubenswrapper[4831]: E1124 08:17:16.894130 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:18 crc kubenswrapper[4831]: I1124 08:17:18.892405 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:18 crc kubenswrapper[4831]: I1124 08:17:18.892427 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:18 crc kubenswrapper[4831]: E1124 08:17:18.892517 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:18 crc kubenswrapper[4831]: I1124 08:17:18.892550 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:18 crc kubenswrapper[4831]: I1124 08:17:18.892585 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:18 crc kubenswrapper[4831]: E1124 08:17:18.892645 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:18 crc kubenswrapper[4831]: E1124 08:17:18.892707 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:18 crc kubenswrapper[4831]: E1124 08:17:18.892759 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:20 crc kubenswrapper[4831]: I1124 08:17:20.893113 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:20 crc kubenswrapper[4831]: I1124 08:17:20.893144 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:20 crc kubenswrapper[4831]: E1124 08:17:20.893253 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:20 crc kubenswrapper[4831]: I1124 08:17:20.893363 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:20 crc kubenswrapper[4831]: I1124 08:17:20.893555 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:20 crc kubenswrapper[4831]: E1124 08:17:20.893653 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:20 crc kubenswrapper[4831]: E1124 08:17:20.893879 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:20 crc kubenswrapper[4831]: E1124 08:17:20.893933 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.405463 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/1.log" Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.406121 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/0.log" Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.406218 4831 generic.go:334] "Generic (PLEG): container finished" podID="89aea50f-21a1-4854-a107-6e38c780166d" containerID="a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d" exitCode=1 Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.406275 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerDied","Data":"a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d"} Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.406381 4831 scope.go:117] "RemoveContainer" containerID="976ea75d9d02047daf62e602ccd24c1f02e1fa389eb32d627c97e9fac7d3fd4b" Nov 24 08:17:21 crc kubenswrapper[4831]: I1124 08:17:21.406830 4831 scope.go:117] "RemoveContainer" containerID="a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d" Nov 24 08:17:21 crc kubenswrapper[4831]: E1124 08:17:21.407003 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-k6hd5_openshift-multus(89aea50f-21a1-4854-a107-6e38c780166d)\"" pod="openshift-multus/multus-k6hd5" podUID="89aea50f-21a1-4854-a107-6e38c780166d" Nov 24 08:17:22 crc kubenswrapper[4831]: I1124 08:17:22.410814 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/1.log" Nov 24 08:17:22 crc kubenswrapper[4831]: I1124 08:17:22.892753 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:22 crc kubenswrapper[4831]: I1124 08:17:22.892784 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:22 crc kubenswrapper[4831]: I1124 08:17:22.892783 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:22 crc kubenswrapper[4831]: E1124 08:17:22.892888 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:22 crc kubenswrapper[4831]: E1124 08:17:22.893052 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:22 crc kubenswrapper[4831]: E1124 08:17:22.893115 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:22 crc kubenswrapper[4831]: I1124 08:17:22.893235 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:22 crc kubenswrapper[4831]: E1124 08:17:22.893288 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:23 crc kubenswrapper[4831]: I1124 08:17:23.893176 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:17:23 crc kubenswrapper[4831]: E1124 08:17:23.893890 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-hxqjw_openshift-ovn-kubernetes(77826b29-ac89-4b2a-afd9-746c0b1c13ff)\"" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" Nov 24 08:17:24 crc kubenswrapper[4831]: I1124 08:17:24.893070 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:24 crc kubenswrapper[4831]: I1124 08:17:24.893107 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:24 crc kubenswrapper[4831]: E1124 08:17:24.893260 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:24 crc kubenswrapper[4831]: E1124 08:17:24.893343 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:24 crc kubenswrapper[4831]: I1124 08:17:24.893533 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:24 crc kubenswrapper[4831]: E1124 08:17:24.893612 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:24 crc kubenswrapper[4831]: I1124 08:17:24.893773 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:24 crc kubenswrapper[4831]: E1124 08:17:24.893851 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:26 crc kubenswrapper[4831]: E1124 08:17:26.863956 4831 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 24 08:17:26 crc kubenswrapper[4831]: I1124 08:17:26.893562 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:26 crc kubenswrapper[4831]: I1124 08:17:26.893620 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:26 crc kubenswrapper[4831]: E1124 08:17:26.895382 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:26 crc kubenswrapper[4831]: I1124 08:17:26.895604 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:26 crc kubenswrapper[4831]: E1124 08:17:26.895954 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:26 crc kubenswrapper[4831]: I1124 08:17:26.896254 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:26 crc kubenswrapper[4831]: E1124 08:17:26.897501 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:26 crc kubenswrapper[4831]: E1124 08:17:26.897572 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:27 crc kubenswrapper[4831]: E1124 08:17:27.008230 4831 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:17:28 crc kubenswrapper[4831]: I1124 08:17:28.892343 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:28 crc kubenswrapper[4831]: I1124 08:17:28.892407 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:28 crc kubenswrapper[4831]: I1124 08:17:28.892357 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:28 crc kubenswrapper[4831]: I1124 08:17:28.892354 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:28 crc kubenswrapper[4831]: E1124 08:17:28.892465 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:28 crc kubenswrapper[4831]: E1124 08:17:28.892534 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:28 crc kubenswrapper[4831]: E1124 08:17:28.892601 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:28 crc kubenswrapper[4831]: E1124 08:17:28.892677 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:30 crc kubenswrapper[4831]: I1124 08:17:30.893512 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:30 crc kubenswrapper[4831]: I1124 08:17:30.893607 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:30 crc kubenswrapper[4831]: E1124 08:17:30.893659 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:30 crc kubenswrapper[4831]: E1124 08:17:30.893792 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:30 crc kubenswrapper[4831]: I1124 08:17:30.893863 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:30 crc kubenswrapper[4831]: E1124 08:17:30.893951 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:30 crc kubenswrapper[4831]: I1124 08:17:30.894041 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:30 crc kubenswrapper[4831]: E1124 08:17:30.894115 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:32 crc kubenswrapper[4831]: E1124 08:17:32.009562 4831 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:17:32 crc kubenswrapper[4831]: I1124 08:17:32.893126 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:32 crc kubenswrapper[4831]: E1124 08:17:32.893305 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:32 crc kubenswrapper[4831]: I1124 08:17:32.893632 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:32 crc kubenswrapper[4831]: I1124 08:17:32.893704 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:32 crc kubenswrapper[4831]: I1124 08:17:32.893659 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:32 crc kubenswrapper[4831]: E1124 08:17:32.893832 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:32 crc kubenswrapper[4831]: E1124 08:17:32.893912 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:32 crc kubenswrapper[4831]: E1124 08:17:32.893986 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.892496 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.892497 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:34 crc kubenswrapper[4831]: E1124 08:17:34.892621 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.892636 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.892954 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:34 crc kubenswrapper[4831]: E1124 08:17:34.893099 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:34 crc kubenswrapper[4831]: E1124 08:17:34.893142 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:34 crc kubenswrapper[4831]: E1124 08:17:34.893382 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.893426 4831 scope.go:117] "RemoveContainer" containerID="a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d" Nov 24 08:17:34 crc kubenswrapper[4831]: I1124 08:17:34.893490 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.458480 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/1.log" Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.458791 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerStarted","Data":"efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac"} Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.460412 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/3.log" Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.462638 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerStarted","Data":"3e033d089aeadb785bd84f8d5fdbeffd332301028b5ac8c474e9500b26d4932c"} Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.463234 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.499882 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podStartSLOduration=109.499866085 podStartE2EDuration="1m49.499866085s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:35.499141604 +0000 UTC m=+129.374286777" watchObservedRunningTime="2025-11-24 08:17:35.499866085 +0000 UTC m=+129.375011258" Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.838363 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k45nx"] Nov 24 08:17:35 crc kubenswrapper[4831]: I1124 08:17:35.838737 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:35 crc kubenswrapper[4831]: E1124 08:17:35.838954 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:36 crc kubenswrapper[4831]: I1124 08:17:36.893273 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:36 crc kubenswrapper[4831]: E1124 08:17:36.894178 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:36 crc kubenswrapper[4831]: I1124 08:17:36.894213 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:36 crc kubenswrapper[4831]: I1124 08:17:36.894219 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:36 crc kubenswrapper[4831]: E1124 08:17:36.894299 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:36 crc kubenswrapper[4831]: E1124 08:17:36.894391 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:37 crc kubenswrapper[4831]: E1124 08:17:37.010343 4831 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 24 08:17:37 crc kubenswrapper[4831]: I1124 08:17:37.892817 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:37 crc kubenswrapper[4831]: E1124 08:17:37.892958 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:38 crc kubenswrapper[4831]: I1124 08:17:38.893060 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:38 crc kubenswrapper[4831]: I1124 08:17:38.893123 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:38 crc kubenswrapper[4831]: I1124 08:17:38.893180 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:38 crc kubenswrapper[4831]: E1124 08:17:38.893219 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:38 crc kubenswrapper[4831]: E1124 08:17:38.893337 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:38 crc kubenswrapper[4831]: E1124 08:17:38.893419 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:39 crc kubenswrapper[4831]: I1124 08:17:39.892726 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:39 crc kubenswrapper[4831]: E1124 08:17:39.892860 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:40 crc kubenswrapper[4831]: I1124 08:17:40.893270 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:40 crc kubenswrapper[4831]: I1124 08:17:40.893273 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:40 crc kubenswrapper[4831]: I1124 08:17:40.893369 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:40 crc kubenswrapper[4831]: E1124 08:17:40.893523 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 24 08:17:40 crc kubenswrapper[4831]: E1124 08:17:40.893858 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 24 08:17:40 crc kubenswrapper[4831]: E1124 08:17:40.893988 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 24 08:17:41 crc kubenswrapper[4831]: I1124 08:17:41.892563 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:41 crc kubenswrapper[4831]: E1124 08:17:41.892686 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-k45nx" podUID="650855b7-20e6-4261-a212-4bc75a671b96" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.892978 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.893497 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.893571 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.895649 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.897347 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.898238 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 24 08:17:42 crc kubenswrapper[4831]: I1124 08:17:42.899307 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 24 08:17:43 crc kubenswrapper[4831]: I1124 08:17:43.892798 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:17:43 crc kubenswrapper[4831]: I1124 08:17:43.896877 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 24 08:17:43 crc kubenswrapper[4831]: I1124 08:17:43.898101 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.008633 4831 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.040037 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.040837 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.041609 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.042124 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.042438 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kzxvh"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.043181 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-z784l"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.043216 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.043661 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.056608 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.057146 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.062745 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.063035 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.063399 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.063585 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.063945 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.064632 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.066552 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.067114 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.068878 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.069378 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.069833 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.070850 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sq6n7"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.072336 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.074891 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.075354 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.078392 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fdjbw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.078820 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-4dbcf"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.079098 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qkrmn"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.079597 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.080019 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.080392 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.080847 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.081163 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.081418 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.081597 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.082439 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.082717 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.088939 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.089575 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cx7sp"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.089947 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.090442 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.092912 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.093612 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.097761 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.101209 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.101534 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.101712 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.102084 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.104533 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.109892 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.109948 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.125948 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.126066 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.126132 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.126176 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.126258 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.126420 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130405 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130462 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130719 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130809 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130920 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130969 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131087 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131105 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131215 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131261 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130970 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131409 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131548 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131579 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130931 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131748 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131793 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.130932 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.131109 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.142304 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.145055 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.147069 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.147831 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.148151 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.148459 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.149236 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.150678 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.152649 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.153624 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.163906 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.163935 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.163960 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.163977 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-service-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.163993 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-images\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164007 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164023 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e476233b-1b31-4866-908b-6ae2a89f390e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164036 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-client\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164050 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh874\" (UniqueName: \"kubernetes.io/projected/6da584cb-b41f-4a9c-8e23-c9ae53240a22-kube-api-access-rh874\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164072 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164087 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsjpz\" (UniqueName: \"kubernetes.io/projected/71daccd0-f1fb-4adf-a125-e5772cb2f759-kube-api-access-rsjpz\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164102 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164117 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164143 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164161 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164175 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164191 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12a303-77ed-4145-b211-12926737b82e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164209 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164226 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-serving-cert\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164241 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4759\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-kube-api-access-n4759\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164254 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-config\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164270 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8nq4\" (UniqueName: \"kubernetes.io/projected/e476233b-1b31-4866-908b-6ae2a89f390e-kube-api-access-c8nq4\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164286 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71daccd0-f1fb-4adf-a125-e5772cb2f759-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164301 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22rzm\" (UniqueName: \"kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164334 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9e12a303-77ed-4145-b211-12926737b82e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164351 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe7bf71-f32f-4394-93cc-4e3157327c5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164389 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lmvt\" (UniqueName: \"kubernetes.io/projected/446d711b-f141-4823-8a50-b7f88ff704ea-kube-api-access-4lmvt\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164406 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e476233b-1b31-4866-908b-6ae2a89f390e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164470 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164486 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlhv7\" (UniqueName: \"kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164501 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446d711b-f141-4823-8a50-b7f88ff704ea-serving-cert\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164515 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-config\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164532 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzvfg\" (UniqueName: \"kubernetes.io/projected/abe7bf71-f32f-4394-93cc-4e3157327c5a-kube-api-access-pzvfg\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164546 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164561 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-config\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164578 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpbhc\" (UniqueName: \"kubernetes.io/projected/9e12a303-77ed-4145-b211-12926737b82e-kube-api-access-xpbhc\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164593 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-trusted-ca\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164609 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164625 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164639 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.164653 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zjfx\" (UniqueName: \"kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.165486 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.166015 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.166190 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.166285 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.166683 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.168224 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.179208 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.179864 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180143 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180298 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180429 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180530 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180627 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180809 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180148 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180540 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.180594 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181497 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181441 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181706 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181810 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181622 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.181670 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182177 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182375 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182496 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182658 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182453 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.182905 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.184386 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.188019 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.188201 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.188418 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.188598 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.189524 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.189577 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.189696 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.189823 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.189892 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190010 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190095 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190405 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190490 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190576 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.190491 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.191727 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.198422 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.206563 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.206936 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.207217 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.207706 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.208024 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.208558 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.210931 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.217586 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.218763 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.218885 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.225542 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.226537 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.227207 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.250391 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.253575 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.255942 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.271081 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.273795 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.273923 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.275773 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.276336 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kzxvh"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.276386 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.276510 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.276773 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pw597"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.276884 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.277035 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.277694 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.277863 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-service-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278000 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e476233b-1b31-4866-908b-6ae2a89f390e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278106 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-images\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278187 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278385 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-client\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278460 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh874\" (UniqueName: \"kubernetes.io/projected/6da584cb-b41f-4a9c-8e23-c9ae53240a22-kube-api-access-rh874\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278529 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsjpz\" (UniqueName: \"kubernetes.io/projected/71daccd0-f1fb-4adf-a125-e5772cb2f759-kube-api-access-rsjpz\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278605 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278713 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278808 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278883 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.278950 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279019 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12a303-77ed-4145-b211-12926737b82e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279093 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279159 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-serving-cert\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279225 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4759\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-kube-api-access-n4759\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279986 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-config\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280071 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22rzm\" (UniqueName: \"kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280144 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-service-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280152 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9e12a303-77ed-4145-b211-12926737b82e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280230 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8nq4\" (UniqueName: \"kubernetes.io/projected/e476233b-1b31-4866-908b-6ae2a89f390e-kube-api-access-c8nq4\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280258 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71daccd0-f1fb-4adf-a125-e5772cb2f759-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280286 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe7bf71-f32f-4394-93cc-4e3157327c5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280348 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280384 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lmvt\" (UniqueName: \"kubernetes.io/projected/446d711b-f141-4823-8a50-b7f88ff704ea-kube-api-access-4lmvt\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280410 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e476233b-1b31-4866-908b-6ae2a89f390e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280433 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280455 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlhv7\" (UniqueName: \"kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280477 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446d711b-f141-4823-8a50-b7f88ff704ea-serving-cert\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280498 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-config\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280524 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzvfg\" (UniqueName: \"kubernetes.io/projected/abe7bf71-f32f-4394-93cc-4e3157327c5a-kube-api-access-pzvfg\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280553 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280575 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-config\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280600 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpbhc\" (UniqueName: \"kubernetes.io/projected/9e12a303-77ed-4145-b211-12926737b82e-kube-api-access-xpbhc\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280627 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280652 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-trusted-ca\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280676 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280709 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280733 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zjfx\" (UniqueName: \"kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280769 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.280797 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.281303 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.281702 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.282077 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9e12a303-77ed-4145-b211-12926737b82e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.282745 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.283856 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.286725 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.287073 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.287625 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.289136 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.279536 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-images\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.289515 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.290157 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abe7bf71-f32f-4394-93cc-4e3157327c5a-config\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.290141 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.290570 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-client\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.291444 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-config\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.291605 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.294731 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ft42l"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.295738 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-prssr"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.296299 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.297151 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.297437 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.298271 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.300880 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/446d711b-f141-4823-8a50-b7f88ff704ea-trusted-ca\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.295358 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.301445 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-config\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.302483 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.302984 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.303883 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.303935 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.304091 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.304962 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.309880 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.311465 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.317180 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.317783 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.318145 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.318580 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.318821 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.319176 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.320796 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6da584cb-b41f-4a9c-8e23-c9ae53240a22-etcd-ca\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.321349 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.321566 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.321603 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e476233b-1b31-4866-908b-6ae2a89f390e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.325564 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446d711b-f141-4823-8a50-b7f88ff704ea-serving-cert\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.326525 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6da584cb-b41f-4a9c-8e23-c9ae53240a22-serving-cert\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.328300 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.329715 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71daccd0-f1fb-4adf-a125-e5772cb2f759-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.330691 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.333906 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e12a303-77ed-4145-b211-12926737b82e-serving-cert\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.339078 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/abe7bf71-f32f-4394-93cc-4e3157327c5a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.339149 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.341865 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e476233b-1b31-4866-908b-6ae2a89f390e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.343096 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.344761 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.348509 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.349128 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.349263 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.349568 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vczfv"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.349606 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.349669 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.350301 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bq4m2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.355190 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.355390 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.350540 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.356585 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.366367 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.368367 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-mn8dg"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.368996 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.369346 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.370461 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qkrmn"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.373656 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4dbcf"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.374484 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.378136 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.378865 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.385120 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sq6n7"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.385163 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mn8dg"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.385173 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.387740 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.394131 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.395892 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cx7sp"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.401818 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.404091 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.404437 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.407286 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pw597"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.408120 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rrssz"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.409685 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.410032 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2ftjk"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.411114 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.411131 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.412566 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.414138 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.417120 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.419196 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fdjbw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.420636 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.423147 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.423996 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.426302 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.428982 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.430727 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rrssz"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.432309 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-prssr"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.433797 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.435221 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vczfv"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.437845 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.439131 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmw8"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.440160 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.441665 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bq4m2"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.442820 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmw8"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.443652 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.444121 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.446237 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.447306 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc"] Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.463022 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.482967 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.510363 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.543162 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.564582 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.584471 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.604133 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.624541 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.644298 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.662874 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.683542 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.703422 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.723507 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.743370 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.763404 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.782953 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.815947 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.823429 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.843430 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.862679 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.882826 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.903013 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.922793 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.943277 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.977681 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsjpz\" (UniqueName: \"kubernetes.io/projected/71daccd0-f1fb-4adf-a125-e5772cb2f759-kube-api-access-rsjpz\") pod \"cluster-samples-operator-665b6dd947-4bw9d\" (UID: \"71daccd0-f1fb-4adf-a125-e5772cb2f759\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.995881 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" Nov 24 08:17:45 crc kubenswrapper[4831]: I1124 08:17:45.996385 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4759\" (UniqueName: \"kubernetes.io/projected/b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5-kube-api-access-n4759\") pod \"cluster-image-registry-operator-dc59b4c8b-5d6v2\" (UID: \"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.017861 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22rzm\" (UniqueName: \"kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm\") pod \"controller-manager-879f6c89f-k2k6z\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.036838 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh874\" (UniqueName: \"kubernetes.io/projected/6da584cb-b41f-4a9c-8e23-c9ae53240a22-kube-api-access-rh874\") pod \"etcd-operator-b45778765-cx7sp\" (UID: \"6da584cb-b41f-4a9c-8e23-c9ae53240a22\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.058560 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzvfg\" (UniqueName: \"kubernetes.io/projected/abe7bf71-f32f-4394-93cc-4e3157327c5a-kube-api-access-pzvfg\") pod \"machine-api-operator-5694c8668f-qkrmn\" (UID: \"abe7bf71-f32f-4394-93cc-4e3157327c5a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.064677 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.084969 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lmvt\" (UniqueName: \"kubernetes.io/projected/446d711b-f141-4823-8a50-b7f88ff704ea-kube-api-access-4lmvt\") pod \"console-operator-58897d9998-sq6n7\" (UID: \"446d711b-f141-4823-8a50-b7f88ff704ea\") " pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.089306 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.103233 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8nq4\" (UniqueName: \"kubernetes.io/projected/e476233b-1b31-4866-908b-6ae2a89f390e-kube-api-access-c8nq4\") pod \"openshift-apiserver-operator-796bbdcf4f-4szpt\" (UID: \"e476233b-1b31-4866-908b-6ae2a89f390e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.104031 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.124170 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.143684 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.159496 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.163913 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.186900 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.204770 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.206606 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.224145 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.244214 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.253494 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.263373 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.270411 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-qkrmn"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.283618 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 24 08:17:46 crc kubenswrapper[4831]: W1124 08:17:46.291983 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabe7bf71_f32f_4394_93cc_4e3157327c5a.slice/crio-2562b323c85898abdf1c333f8f3f7d403d23840c927957003e119993866d8f1d WatchSource:0}: Error finding container 2562b323c85898abdf1c333f8f3f7d403d23840c927957003e119993866d8f1d: Status 404 returned error can't find the container with id 2562b323c85898abdf1c333f8f3f7d403d23840c927957003e119993866d8f1d Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.301972 4831 request.go:700] Waited for 1.002507952s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.303754 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.306686 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.320578 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.324310 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.329701 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.371794 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cx7sp"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.386128 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zjfx\" (UniqueName: \"kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx\") pod \"console-f9d7485db-9kqsc\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.392450 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpbhc\" (UniqueName: \"kubernetes.io/projected/9e12a303-77ed-4145-b211-12926737b82e-kube-api-access-xpbhc\") pod \"openshift-config-operator-7777fb866f-cc7tj\" (UID: \"9e12a303-77ed-4145-b211-12926737b82e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.393410 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: W1124 08:17:46.399014 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6da584cb_b41f_4a9c_8e23_c9ae53240a22.slice/crio-33865367b304655dafa774052abf6ffd22da7bc8e18c29bd2c39e09160451712 WatchSource:0}: Error finding container 33865367b304655dafa774052abf6ffd22da7bc8e18c29bd2c39e09160451712: Status 404 returned error can't find the container with id 33865367b304655dafa774052abf6ffd22da7bc8e18c29bd2c39e09160451712 Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.407396 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.423664 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.448641 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.464612 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.481230 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.483601 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.498963 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" event={"ID":"abe7bf71-f32f-4394-93cc-4e3157327c5a","Type":"ContainerStarted","Data":"2562b323c85898abdf1c333f8f3f7d403d23840c927957003e119993866d8f1d"} Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.499981 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" event={"ID":"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5","Type":"ContainerStarted","Data":"2de76769e1a21e3b362d7b2c15eb5a775614b6e705663fd4e53540f0698b2e34"} Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.501980 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" event={"ID":"71daccd0-f1fb-4adf-a125-e5772cb2f759","Type":"ContainerStarted","Data":"7437f1b39269373774c19cbf20c675feabc341abdf03edc72d8a7b0418ce5c0f"} Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.503080 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" event={"ID":"6da584cb-b41f-4a9c-8e23-c9ae53240a22","Type":"ContainerStarted","Data":"33865367b304655dafa774052abf6ffd22da7bc8e18c29bd2c39e09160451712"} Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.503204 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.516763 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.524120 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.537332 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.543866 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.562952 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.583416 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sq6n7"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.598308 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlhv7\" (UniqueName: \"kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7\") pod \"route-controller-manager-6576b87f9c-6lg2c\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.609879 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.623471 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.631552 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.643433 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.656069 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.663239 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 24 08:17:46 crc kubenswrapper[4831]: W1124 08:17:46.683530 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode476233b_1b31_4866_908b_6ae2a89f390e.slice/crio-74934f4e5be0147296bbcf382e2d50939688d8ca992081426686ba35f2b643a6 WatchSource:0}: Error finding container 74934f4e5be0147296bbcf382e2d50939688d8ca992081426686ba35f2b643a6: Status 404 returned error can't find the container with id 74934f4e5be0147296bbcf382e2d50939688d8ca992081426686ba35f2b643a6 Nov 24 08:17:46 crc kubenswrapper[4831]: W1124 08:17:46.692372 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod446d711b_f141_4823_8a50_b7f88ff704ea.slice/crio-8260be041a9faf511768e0bbc93c2a0110ed44f2584886a1f268581b9e2b2f25 WatchSource:0}: Error finding container 8260be041a9faf511768e0bbc93c2a0110ed44f2584886a1f268581b9e2b2f25: Status 404 returned error can't find the container with id 8260be041a9faf511768e0bbc93c2a0110ed44f2584886a1f268581b9e2b2f25 Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.704220 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.713899 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.713936 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.713987 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-client\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714004 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714020 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-service-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714036 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-encryption-config\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714053 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714068 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/df388cd7-d680-494b-b6f8-d1e6c51cd06b-machine-approver-tls\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714085 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714104 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkm25\" (UniqueName: \"kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714125 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-serving-cert\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714147 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714166 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-serving-cert\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714185 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgm27\" (UniqueName: \"kubernetes.io/projected/984066f7-5187-41f4-96b8-9cacec55a5d3-kube-api-access-lgm27\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714206 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjw94\" (UniqueName: \"kubernetes.io/projected/df388cd7-d680-494b-b6f8-d1e6c51cd06b-kube-api-access-gjw94\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714231 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714252 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-audit\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714274 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714295 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-auth-proxy-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714330 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714351 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-dir\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714378 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8mvr\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714396 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714410 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-node-pullsecrets\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714429 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714446 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714471 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714524 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-client\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714540 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714557 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714581 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714599 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714738 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714783 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-encryption-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714811 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714828 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714849 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-audit-dir\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714876 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714900 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: E1124 08:17:46.714924 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.214909557 +0000 UTC m=+141.090054790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714945 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714968 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.714990 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-image-import-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715004 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-serving-cert\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715020 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppxnp\" (UniqueName: \"kubernetes.io/projected/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-kube-api-access-ppxnp\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715045 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715060 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-policies\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715076 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgn5v\" (UniqueName: \"kubernetes.io/projected/02514c19-c154-4055-ae24-eb2acaeb03d3-kube-api-access-wgn5v\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715095 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55x5q\" (UniqueName: \"kubernetes.io/projected/4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52-kube-api-access-55x5q\") pod \"downloads-7954f5f757-4dbcf\" (UID: \"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52\") " pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715118 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-config\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715137 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.715157 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.723097 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.744124 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.763636 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.816304 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.816609 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.817504 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818064 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-policies\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818101 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df09a1f6-dfe4-4fd2-94a0-280c6a701737-proxy-tls\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818124 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7q7q\" (UniqueName: \"kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818159 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/972d8e1c-6711-4b23-8014-8c463852b780-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818195 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55x5q\" (UniqueName: \"kubernetes.io/projected/4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52-kube-api-access-55x5q\") pod \"downloads-7954f5f757-4dbcf\" (UID: \"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52\") " pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818217 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dqwb\" (UniqueName: \"kubernetes.io/projected/ef536ac2-f27c-47f3-9571-b69f72d65178-kube-api-access-6dqwb\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818241 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmvs9\" (UniqueName: \"kubernetes.io/projected/3aba0829-409c-4395-bf72-9eba0a55c877-kube-api-access-cmvs9\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818332 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-srv-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818355 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818382 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818405 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktnxp\" (UniqueName: \"kubernetes.io/projected/34330ad7-9d56-4d26-bf45-15f34990ea8b-kube-api-access-ktnxp\") pod \"migrator-59844c95c7-zxwd2\" (UID: \"34330ad7-9d56-4d26-bf45-15f34990ea8b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818460 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818481 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd58485-1abf-48ee-b3d7-50ed174205a3-metrics-tls\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818633 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-certs\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818706 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818734 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818805 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fba95cbe-ea61-442e-b35a-fc31277e4e7e-tmpfs\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818846 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818883 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/df388cd7-d680-494b-b6f8-d1e6c51cd06b-machine-approver-tls\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818903 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-encryption-config\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818944 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef536ac2-f27c-47f3-9571-b69f72d65178-serving-cert\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.818966 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b6963b-9605-46c1-b800-280539eec365-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.819031 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-serving-cert\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.819054 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.819952 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-serving-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: E1124 08:17:46.820250 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.319784339 +0000 UTC m=+141.194929612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820291 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820385 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-cert\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820421 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-audit\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820466 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820504 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.820532 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.821744 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-policies\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.823067 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.823925 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824583 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-audit\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824618 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824785 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824920 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-auth-proxy-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824942 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-dir\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.824964 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fcd58485-1abf-48ee-b3d7-50ed174205a3-trusted-ca\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825012 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p47w\" (UniqueName: \"kubernetes.io/projected/c6ffcfa0-86f7-41ca-934f-a633c2450a32-kube-api-access-2p47w\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825031 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972d8e1c-6711-4b23-8014-8c463852b780-config\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825048 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f83accac-e306-4997-aba7-9276b9f9409c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825070 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-key\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825087 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef536ac2-f27c-47f3-9571-b69f72d65178-config\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825102 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cb983bd-e989-4487-b31a-c0328ca2960f-service-ca-bundle\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825120 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-metrics-certs\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825138 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-registration-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825154 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlqv5\" (UniqueName: \"kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825212 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825231 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825287 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825298 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-client\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825356 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825401 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/02514c19-c154-4055-ae24-eb2acaeb03d3-audit-dir\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.825355 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kfcr\" (UniqueName: \"kubernetes.io/projected/0cb983bd-e989-4487-b31a-c0328ca2960f-kube-api-access-8kfcr\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.829749 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.829803 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.829844 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830275 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830546 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8khb\" (UniqueName: \"kubernetes.io/projected/d4f0d807-8634-4c21-a4b8-924ad3463b0c-kube-api-access-t8khb\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830579 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpkbx\" (UniqueName: \"kubernetes.io/projected/f83accac-e306-4997-aba7-9276b9f9409c-kube-api-access-rpkbx\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830627 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-mountpoint-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830660 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7ce732b-7d7c-43f9-8944-72bd510a3be2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830705 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h456\" (UniqueName: \"kubernetes.io/projected/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-kube-api-access-7h456\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.830734 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.831012 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/df388cd7-d680-494b-b6f8-d1e6c51cd06b-machine-approver-tls\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.831493 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.831855 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-encryption-config\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.833030 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-serving-cert\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.834689 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.835059 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-audit-dir\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.835128 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ce732b-7d7c-43f9-8944-72bd510a3be2-config\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836313 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836587 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-stats-auth\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836649 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwnkc\" (UniqueName: \"kubernetes.io/projected/ce29f16d-07cb-4d33-95bd-035d6d012ad2-kube-api-access-fwnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836713 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f2cd\" (UniqueName: \"kubernetes.io/projected/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-kube-api-access-7f2cd\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836757 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-audit-dir\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836809 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.836873 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-images\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837253 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837370 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837405 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twds5\" (UniqueName: \"kubernetes.io/projected/fba95cbe-ea61-442e-b35a-fc31277e4e7e-kube-api-access-twds5\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837462 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b6963b-9605-46c1-b800-280539eec365-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837493 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972d8e1c-6711-4b23-8014-8c463852b780-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837514 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df388cd7-d680-494b-b6f8-d1e6c51cd06b-auth-proxy-config\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837522 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce29f16d-07cb-4d33-95bd-035d6d012ad2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837686 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.837790 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-image-import-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.838487 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839387 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgn5v\" (UniqueName: \"kubernetes.io/projected/02514c19-c154-4055-ae24-eb2acaeb03d3-kube-api-access-wgn5v\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839610 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-default-certificate\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839694 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/984066f7-5187-41f4-96b8-9cacec55a5d3-image-import-ca\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839820 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-config\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839873 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ce732b-7d7c-43f9-8944-72bd510a3be2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.839908 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.840467 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-config\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.840734 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.840923 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/02514c19-c154-4055-ae24-eb2acaeb03d3-etcd-client\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841059 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-client\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841085 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841125 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd19c608-b11b-4ef4-9efa-11710fec5a1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841148 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-service-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841176 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-csi-data-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841202 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841262 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxxnp\" (UniqueName: \"kubernetes.io/projected/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-kube-api-access-lxxnp\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841283 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd19c608-b11b-4ef4-9efa-11710fec5a1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841306 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59d6b\" (UniqueName: \"kubernetes.io/projected/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-kube-api-access-59d6b\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841353 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzwbz\" (UniqueName: \"kubernetes.io/projected/69b6963b-9605-46c1-b800-280539eec365-kube-api-access-qzwbz\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jvpk\" (UniqueName: \"kubernetes.io/projected/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-kube-api-access-9jvpk\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841442 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74tvz\" (UniqueName: \"kubernetes.io/projected/df09a1f6-dfe4-4fd2-94a0-280c6a701737-kube-api-access-74tvz\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841467 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841485 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkm25\" (UniqueName: \"kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841508 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-plugins-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841541 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841562 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-serving-cert\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841579 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgm27\" (UniqueName: \"kubernetes.io/projected/984066f7-5187-41f4-96b8-9cacec55a5d3-kube-api-access-lgm27\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841599 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjw94\" (UniqueName: \"kubernetes.io/projected/df388cd7-d680-494b-b6f8-d1e6c51cd06b-kube-api-access-gjw94\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841440 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841691 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c6ffcfa0-86f7-41ca-934f-a633c2450a32-metrics-tls\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841741 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841772 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3aba0829-409c-4395-bf72-9eba0a55c877-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841795 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841816 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-node-bootstrap-token\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841832 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxs8k\" (UniqueName: \"kubernetes.io/projected/1a934030-84e1-45c1-a58f-00b41a6d1b38-kube-api-access-qxs8k\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.841850 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-webhook-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.843049 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.843597 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.843667 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02514c19-c154-4055-ae24-eb2acaeb03d3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.845698 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-service-ca-bundle\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.845780 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8mvr\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.845825 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9gml\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-kube-api-access-p9gml\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.845901 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.845969 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-node-pullsecrets\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846042 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846132 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd19c608-b11b-4ef4-9efa-11710fec5a1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846187 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846216 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3aba0829-409c-4395-bf72-9eba0a55c877-proxy-tls\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846265 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846294 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846340 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846368 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6ffcfa0-86f7-41ca-934f-a633c2450a32-config-volume\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846510 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846539 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-socket-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846583 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-encryption-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846612 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846642 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846671 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846695 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt86t\" (UniqueName: \"kubernetes.io/projected/e5984c81-33f3-47e5-9911-2c2b18f07f23-kube-api-access-dt86t\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846742 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl9wr\" (UniqueName: \"kubernetes.io/projected/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-kube-api-access-kl9wr\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846889 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-metrics-tls\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846927 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.846953 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-srv-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.847004 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-serving-cert\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.847033 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppxnp\" (UniqueName: \"kubernetes.io/projected/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-kube-api-access-ppxnp\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.847040 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.847067 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-cabundle\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.847772 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.848745 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-etcd-client\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.849064 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/984066f7-5187-41f4-96b8-9cacec55a5d3-node-pullsecrets\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.850690 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: E1124 08:17:46.851208 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.351184721 +0000 UTC m=+141.226329874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.851742 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.851803 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.854736 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.856853 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.858146 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-serving-cert\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.859216 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.860823 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.862370 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.862993 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.863616 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.877441 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-encryption-config\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.877772 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/984066f7-5187-41f4-96b8-9cacec55a5d3-serving-cert\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.900041 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.902571 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.923585 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.944056 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.947937 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948070 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948096 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd19c608-b11b-4ef4-9efa-11710fec5a1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948115 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxxnp\" (UniqueName: \"kubernetes.io/projected/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-kube-api-access-lxxnp\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948130 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd19c608-b11b-4ef4-9efa-11710fec5a1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948145 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-csi-data-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948161 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948176 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzwbz\" (UniqueName: \"kubernetes.io/projected/69b6963b-9605-46c1-b800-280539eec365-kube-api-access-qzwbz\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948191 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jvpk\" (UniqueName: \"kubernetes.io/projected/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-kube-api-access-9jvpk\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948206 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59d6b\" (UniqueName: \"kubernetes.io/projected/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-kube-api-access-59d6b\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948222 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74tvz\" (UniqueName: \"kubernetes.io/projected/df09a1f6-dfe4-4fd2-94a0-280c6a701737-kube-api-access-74tvz\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948251 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-plugins-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948268 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948310 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c6ffcfa0-86f7-41ca-934f-a633c2450a32-metrics-tls\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948368 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3aba0829-409c-4395-bf72-9eba0a55c877-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948383 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-node-bootstrap-token\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948399 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxs8k\" (UniqueName: \"kubernetes.io/projected/1a934030-84e1-45c1-a58f-00b41a6d1b38-kube-api-access-qxs8k\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948414 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-webhook-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948437 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9gml\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-kube-api-access-p9gml\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948452 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd19c608-b11b-4ef4-9efa-11710fec5a1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948475 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3aba0829-409c-4395-bf72-9eba0a55c877-proxy-tls\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948490 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948503 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6ffcfa0-86f7-41ca-934f-a633c2450a32-config-volume\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948524 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-socket-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948540 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948556 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt86t\" (UniqueName: \"kubernetes.io/projected/e5984c81-33f3-47e5-9911-2c2b18f07f23-kube-api-access-dt86t\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948574 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl9wr\" (UniqueName: \"kubernetes.io/projected/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-kube-api-access-kl9wr\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948588 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-metrics-tls\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948602 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-srv-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948629 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-cabundle\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948642 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df09a1f6-dfe4-4fd2-94a0-280c6a701737-proxy-tls\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948657 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7q7q\" (UniqueName: \"kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948675 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/972d8e1c-6711-4b23-8014-8c463852b780-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948690 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmvs9\" (UniqueName: \"kubernetes.io/projected/3aba0829-409c-4395-bf72-9eba0a55c877-kube-api-access-cmvs9\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948718 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dqwb\" (UniqueName: \"kubernetes.io/projected/ef536ac2-f27c-47f3-9571-b69f72d65178-kube-api-access-6dqwb\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948732 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-srv-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948752 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948767 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktnxp\" (UniqueName: \"kubernetes.io/projected/34330ad7-9d56-4d26-bf45-15f34990ea8b-kube-api-access-ktnxp\") pod \"migrator-59844c95c7-zxwd2\" (UID: \"34330ad7-9d56-4d26-bf45-15f34990ea8b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948782 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd58485-1abf-48ee-b3d7-50ed174205a3-metrics-tls\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948795 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-certs\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948817 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fba95cbe-ea61-442e-b35a-fc31277e4e7e-tmpfs\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948833 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948848 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948862 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef536ac2-f27c-47f3-9571-b69f72d65178-serving-cert\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948876 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b6963b-9605-46c1-b800-280539eec365-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948891 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-cert\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948906 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948921 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fcd58485-1abf-48ee-b3d7-50ed174205a3-trusted-ca\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948937 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p47w\" (UniqueName: \"kubernetes.io/projected/c6ffcfa0-86f7-41ca-934f-a633c2450a32-kube-api-access-2p47w\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948950 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-key\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948964 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972d8e1c-6711-4b23-8014-8c463852b780-config\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948981 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f83accac-e306-4997-aba7-9276b9f9409c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.948995 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef536ac2-f27c-47f3-9571-b69f72d65178-config\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949010 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cb983bd-e989-4487-b31a-c0328ca2960f-service-ca-bundle\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949023 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-metrics-certs\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949038 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-registration-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949053 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlqv5\" (UniqueName: \"kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949068 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kfcr\" (UniqueName: \"kubernetes.io/projected/0cb983bd-e989-4487-b31a-c0328ca2960f-kube-api-access-8kfcr\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949085 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949100 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949114 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8khb\" (UniqueName: \"kubernetes.io/projected/d4f0d807-8634-4c21-a4b8-924ad3463b0c-kube-api-access-t8khb\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949130 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpkbx\" (UniqueName: \"kubernetes.io/projected/f83accac-e306-4997-aba7-9276b9f9409c-kube-api-access-rpkbx\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949149 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h456\" (UniqueName: \"kubernetes.io/projected/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-kube-api-access-7h456\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949164 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-mountpoint-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949179 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7ce732b-7d7c-43f9-8944-72bd510a3be2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949194 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949210 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ce732b-7d7c-43f9-8944-72bd510a3be2-config\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949225 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-stats-auth\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949240 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-images\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949257 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwnkc\" (UniqueName: \"kubernetes.io/projected/ce29f16d-07cb-4d33-95bd-035d6d012ad2-kube-api-access-fwnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949272 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f2cd\" (UniqueName: \"kubernetes.io/projected/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-kube-api-access-7f2cd\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949288 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twds5\" (UniqueName: \"kubernetes.io/projected/fba95cbe-ea61-442e-b35a-fc31277e4e7e-kube-api-access-twds5\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949310 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b6963b-9605-46c1-b800-280539eec365-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949355 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972d8e1c-6711-4b23-8014-8c463852b780-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949373 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce29f16d-07cb-4d33-95bd-035d6d012ad2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949402 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-default-certificate\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.949418 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ce732b-7d7c-43f9-8944-72bd510a3be2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: E1124 08:17:46.950692 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.450678829 +0000 UTC m=+141.325823972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.953652 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7ce732b-7d7c-43f9-8944-72bd510a3be2-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.954944 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.957174 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/972d8e1c-6711-4b23-8014-8c463852b780-config\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.957674 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fcd58485-1abf-48ee-b3d7-50ed174205a3-trusted-ca\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.959049 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-csi-data-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.959454 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fba95cbe-ea61-442e-b35a-fc31277e4e7e-tmpfs\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.960135 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef536ac2-f27c-47f3-9571-b69f72d65178-config\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.960982 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cb983bd-e989-4487-b31a-c0328ca2960f-service-ca-bundle\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.961846 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd19c608-b11b-4ef4-9efa-11710fec5a1a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.962115 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-registration-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.963297 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.963481 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-socket-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.963737 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.964266 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-images\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.964962 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69b6963b-9605-46c1-b800-280539eec365-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.965514 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-plugins-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.970853 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1a934030-84e1-45c1-a58f-00b41a6d1b38-mountpoint-dir\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.971373 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df09a1f6-dfe4-4fd2-94a0-280c6a701737-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.971788 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7ce732b-7d7c-43f9-8944-72bd510a3be2-config\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.973393 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f83accac-e306-4997-aba7-9276b9f9409c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.982200 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/69b6963b-9605-46c1-b800-280539eec365-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.982538 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.982585 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3aba0829-409c-4395-bf72-9eba0a55c877-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.982947 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef536ac2-f27c-47f3-9571-b69f72d65178-serving-cert\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.986260 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.986800 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.986940 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-apiservice-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.987632 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-stats-auth\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.988297 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd19c608-b11b-4ef4-9efa-11710fec5a1a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.988545 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-metrics-certs\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.988946 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce29f16d-07cb-4d33-95bd-035d6d012ad2-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.989225 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df09a1f6-dfe4-4fd2-94a0-280c6a701737-proxy-tls\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.989489 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fcd58485-1abf-48ee-b3d7-50ed174205a3-metrics-tls\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.989786 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.990181 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0cb983bd-e989-4487-b31a-c0328ca2960f-default-certificate\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.990875 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.990896 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.991471 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/972d8e1c-6711-4b23-8014-8c463852b780-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.992200 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-profile-collector-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:46 crc kubenswrapper[4831]: I1124 08:17:46.994175 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.001183 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.001809 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3aba0829-409c-4395-bf72-9eba0a55c877-proxy-tls\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.003073 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fba95cbe-ea61-442e-b35a-fc31277e4e7e-webhook-cert\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.003147 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-srv-cert\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.008992 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.009109 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4f0d807-8634-4c21-a4b8-924ad3463b0c-srv-cert\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.013542 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-metrics-tls\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.028009 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.038379 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-key\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.043863 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.050601 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.050887 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.550876426 +0000 UTC m=+141.426021579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.055522 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-signing-cabundle\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.070426 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.089669 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.104809 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.123167 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.130476 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.134040 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-cert\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:47 crc kubenswrapper[4831]: W1124 08:17:47.141607 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b4b7eb7_c3ef_4cd2_9d28_6b2f4cbc8c3b.slice/crio-4000e3f846b359b1894a560131b349ebab7fde202cbc5b90aa343fd791c45e1c WatchSource:0}: Error finding container 4000e3f846b359b1894a560131b349ebab7fde202cbc5b90aa343fd791c45e1c: Status 404 returned error can't find the container with id 4000e3f846b359b1894a560131b349ebab7fde202cbc5b90aa343fd791c45e1c Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.145492 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.151901 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.152567 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.652543396 +0000 UTC m=+141.527688549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.168546 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.174890 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c6ffcfa0-86f7-41ca-934f-a633c2450a32-config-volume\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.183816 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.203904 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.218394 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c6ffcfa0-86f7-41ca-934f-a633c2450a32-metrics-tls\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.223515 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.243285 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.253619 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.253930 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.753915778 +0000 UTC m=+141.629060931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.256900 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-node-bootstrap-token\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.263489 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.272398 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e5984c81-33f3-47e5-9911-2c2b18f07f23-certs\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.283750 4831 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.303032 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.321994 4831 request.go:700] Waited for 1.881548718s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.323798 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.355217 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.355751 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.855736592 +0000 UTC m=+141.730881745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.383085 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55x5q\" (UniqueName: \"kubernetes.io/projected/4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52-kube-api-access-55x5q\") pod \"downloads-7954f5f757-4dbcf\" (UID: \"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52\") " pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.398844 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.420875 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgn5v\" (UniqueName: \"kubernetes.io/projected/02514c19-c154-4055-ae24-eb2acaeb03d3-kube-api-access-wgn5v\") pod \"apiserver-7bbb656c7d-cdqhk\" (UID: \"02514c19-c154-4055-ae24-eb2acaeb03d3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.440465 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkm25\" (UniqueName: \"kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25\") pod \"oauth-openshift-558db77b4-4nv2q\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.457358 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.457786 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:47.957768253 +0000 UTC m=+141.832913406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.459124 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgm27\" (UniqueName: \"kubernetes.io/projected/984066f7-5187-41f4-96b8-9cacec55a5d3-kube-api-access-lgm27\") pod \"apiserver-76f77b778f-kzxvh\" (UID: \"984066f7-5187-41f4-96b8-9cacec55a5d3\") " pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.469878 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.479383 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjw94\" (UniqueName: \"kubernetes.io/projected/df388cd7-d680-494b-b6f8-d1e6c51cd06b-kube-api-access-gjw94\") pod \"machine-approver-56656f9798-z784l\" (UID: \"df388cd7-d680-494b-b6f8-d1e6c51cd06b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.485266 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.500426 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppxnp\" (UniqueName: \"kubernetes.io/projected/fcc50e57-65ea-4a4a-aeb6-524b44ff9bed-kube-api-access-ppxnp\") pod \"authentication-operator-69f744f599-fdjbw\" (UID: \"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.513232 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" event={"ID":"e476233b-1b31-4866-908b-6ae2a89f390e","Type":"ContainerStarted","Data":"7a223194d5837aaee309981972e6a910fc1d29368c61bed34583e410536242b9"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.513281 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" event={"ID":"e476233b-1b31-4866-908b-6ae2a89f390e","Type":"ContainerStarted","Data":"74934f4e5be0147296bbcf382e2d50939688d8ca992081426686ba35f2b643a6"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.515032 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" event={"ID":"6a33ab7d-6562-4fbe-8f01-ed62511ff801","Type":"ContainerStarted","Data":"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.515079 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" event={"ID":"6a33ab7d-6562-4fbe-8f01-ed62511ff801","Type":"ContainerStarted","Data":"b2cad2a6c46b507020f995c25baaf86bdd3e5867e4c5abc826c9c5535ac679c6"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.515286 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.516266 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" event={"ID":"446d711b-f141-4823-8a50-b7f88ff704ea","Type":"ContainerStarted","Data":"a17ebc415ea29ba19973c4a951c2444025e7d854ffc5404cdd67def5f3d3c023"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.516295 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" event={"ID":"446d711b-f141-4823-8a50-b7f88ff704ea","Type":"ContainerStarted","Data":"8260be041a9faf511768e0bbc93c2a0110ed44f2584886a1f268581b9e2b2f25"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.517050 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.518927 4831 patch_prober.go:28] interesting pod/console-operator-58897d9998-sq6n7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.518982 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" podUID="446d711b-f141-4823-8a50-b7f88ff704ea" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.519759 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" event={"ID":"6da584cb-b41f-4a9c-8e23-c9ae53240a22","Type":"ContainerStarted","Data":"ac9c9e2cc8311e0ec1a5b13eb030c0b9c735618f51d6f031607eefc819b79c12"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.519917 4831 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-k2k6z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.519960 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.521227 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kqsc" event={"ID":"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b","Type":"ContainerStarted","Data":"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.521252 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kqsc" event={"ID":"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b","Type":"ContainerStarted","Data":"cd92280e88715f8cdfd0ca91cab970532d70d929d52cb4011cfde668cd1b2c49"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.523118 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8mvr\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.530181 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" event={"ID":"abe7bf71-f32f-4394-93cc-4e3157327c5a","Type":"ContainerStarted","Data":"2f14aa742dc28465c4f2c70c0473986762442ca34de7e49106d95a1faf0ddbf7"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.530229 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" event={"ID":"abe7bf71-f32f-4394-93cc-4e3157327c5a","Type":"ContainerStarted","Data":"70e0f2a34b335a49d0dc2742143907357ce5e99bababb6ae988717b9c2ec3d4b"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.532546 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" event={"ID":"71daccd0-f1fb-4adf-a125-e5772cb2f759","Type":"ContainerStarted","Data":"aecfa9b13eb74c275ad570e99e9144412505abe97cfbbc1d9df3090520ad6c4f"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.532638 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" event={"ID":"71daccd0-f1fb-4adf-a125-e5772cb2f759","Type":"ContainerStarted","Data":"9df4379eec886b6752969167f1399c40a90105a46058db94e6883e0533885961"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.538777 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" event={"ID":"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b","Type":"ContainerStarted","Data":"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.538840 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" event={"ID":"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b","Type":"ContainerStarted","Data":"4000e3f846b359b1894a560131b349ebab7fde202cbc5b90aa343fd791c45e1c"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.539306 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.540531 4831 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6lg2c container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.540566 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.541906 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktnxp\" (UniqueName: \"kubernetes.io/projected/34330ad7-9d56-4d26-bf45-15f34990ea8b-kube-api-access-ktnxp\") pod \"migrator-59844c95c7-zxwd2\" (UID: \"34330ad7-9d56-4d26-bf45-15f34990ea8b\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.551434 4831 generic.go:334] "Generic (PLEG): container finished" podID="9e12a303-77ed-4145-b211-12926737b82e" containerID="a4787904f579bcc9e904667514d5b4233a9f969c1e1512cf0d0891b6ac81cf19" exitCode=0 Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.551555 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" event={"ID":"9e12a303-77ed-4145-b211-12926737b82e","Type":"ContainerDied","Data":"a4787904f579bcc9e904667514d5b4233a9f969c1e1512cf0d0891b6ac81cf19"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.551607 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" event={"ID":"9e12a303-77ed-4145-b211-12926737b82e","Type":"ContainerStarted","Data":"b6e88057fc6d887fb1adc0b16f8630f2370bb87fd9f0e0b4abf60ed6f50ba6e7"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.554368 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" event={"ID":"b7d5580e-0ef4-4da1-8ed5-2e98cfd6a3c5","Type":"ContainerStarted","Data":"ca27b7cd5dd633c2b9d087cf7ce5d3afa1e9f9e3fe9079bda54fded0c6174271"} Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.559925 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.560184 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.060146423 +0000 UTC m=+141.935291576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.567759 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.569197 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.570881 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dd19c608-b11b-4ef4-9efa-11710fec5a1a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-clzxl\" (UID: \"dd19c608-b11b-4ef4-9efa-11710fec5a1a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.573996 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.07397904 +0000 UTC m=+141.949124193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.575093 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.582452 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.589727 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxxnp\" (UniqueName: \"kubernetes.io/projected/8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9-kube-api-access-lxxnp\") pod \"service-ca-9c57cc56f-vczfv\" (UID: \"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9\") " pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.597333 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.610521 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kfcr\" (UniqueName: \"kubernetes.io/projected/0cb983bd-e989-4487-b31a-c0328ca2960f-kube-api-access-8kfcr\") pod \"router-default-5444994796-ft42l\" (UID: \"0cb983bd-e989-4487-b31a-c0328ca2960f\") " pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.631018 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.641665 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p47w\" (UniqueName: \"kubernetes.io/projected/c6ffcfa0-86f7-41ca-934f-a633c2450a32-kube-api-access-2p47w\") pod \"dns-default-rrssz\" (UID: \"c6ffcfa0-86f7-41ca-934f-a633c2450a32\") " pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.655685 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlqv5\" (UniqueName: \"kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5\") pod \"marketplace-operator-79b997595-5t9z8\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.657242 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.660123 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmvs9\" (UniqueName: \"kubernetes.io/projected/3aba0829-409c-4395-bf72-9eba0a55c877-kube-api-access-cmvs9\") pod \"machine-config-controller-84d6567774-prssr\" (UID: \"3aba0829-409c-4395-bf72-9eba0a55c877\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.673609 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.673929 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.173851739 +0000 UTC m=+142.048996892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.674396 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.682710 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.182692282 +0000 UTC m=+142.057837435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.686855 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7q7q\" (UniqueName: \"kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q\") pod \"collect-profiles-29399535-fpbbw\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.704882 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/972d8e1c-6711-4b23-8014-8c463852b780-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-v29kz\" (UID: \"972d8e1c-6711-4b23-8014-8c463852b780\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.754817 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwnkc\" (UniqueName: \"kubernetes.io/projected/ce29f16d-07cb-4d33-95bd-035d6d012ad2-kube-api-access-fwnkc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rxwsw\" (UID: \"ce29f16d-07cb-4d33-95bd-035d6d012ad2\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.772278 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.773445 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f2cd\" (UniqueName: \"kubernetes.io/projected/f1162ff1-0f44-4168-bcf6-945e5ef9adfd-kube-api-access-7f2cd\") pod \"kube-storage-version-migrator-operator-b67b599dd-nrcst\" (UID: \"f1162ff1-0f44-4168-bcf6-945e5ef9adfd\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.776660 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.776896 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twds5\" (UniqueName: \"kubernetes.io/projected/fba95cbe-ea61-442e-b35a-fc31277e4e7e-kube-api-access-twds5\") pod \"packageserver-d55dfcdfc-d8l7s\" (UID: \"fba95cbe-ea61-442e-b35a-fc31277e4e7e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.777900 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.778268 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.278220946 +0000 UTC m=+142.153366099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.778518 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.779297 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.279285047 +0000 UTC m=+142.154430200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.779873 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.795978 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8khb\" (UniqueName: \"kubernetes.io/projected/d4f0d807-8634-4c21-a4b8-924ad3463b0c-kube-api-access-t8khb\") pod \"catalog-operator-68c6474976-jkrqt\" (UID: \"d4f0d807-8634-4c21-a4b8-924ad3463b0c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.807698 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.815680 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzwbz\" (UniqueName: \"kubernetes.io/projected/69b6963b-9605-46c1-b800-280539eec365-kube-api-access-qzwbz\") pod \"openshift-controller-manager-operator-756b6f6bc6-546qh\" (UID: \"69b6963b-9605-46c1-b800-280539eec365\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.815969 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.832634 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.841122 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jvpk\" (UniqueName: \"kubernetes.io/projected/8d7c3f8f-edbc-449a-b891-3c91952dfa2f-kube-api-access-9jvpk\") pod \"multus-admission-controller-857f4d67dd-pw597\" (UID: \"8d7c3f8f-edbc-449a-b891-3c91952dfa2f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.844194 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59d6b\" (UniqueName: \"kubernetes.io/projected/3fab9f8d-f413-444f-b32b-d3f09cc3b2b8-kube-api-access-59d6b\") pod \"dns-operator-744455d44c-bq4m2\" (UID: \"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8\") " pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.851779 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.864496 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74tvz\" (UniqueName: \"kubernetes.io/projected/df09a1f6-dfe4-4fd2-94a0-280c6a701737-kube-api-access-74tvz\") pod \"machine-config-operator-74547568cd-hwbrt\" (UID: \"df09a1f6-dfe4-4fd2-94a0-280c6a701737\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.864704 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.881890 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.882564 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.382547602 +0000 UTC m=+142.257692755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.882829 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.893139 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.894721 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.919240 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.921785 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.927122 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt86t\" (UniqueName: \"kubernetes.io/projected/e5984c81-33f3-47e5-9911-2c2b18f07f23-kube-api-access-dt86t\") pod \"machine-config-server-2ftjk\" (UID: \"e5984c81-33f3-47e5-9911-2c2b18f07f23\") " pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.943585 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpkbx\" (UniqueName: \"kubernetes.io/projected/f83accac-e306-4997-aba7-9276b9f9409c-kube-api-access-rpkbx\") pod \"package-server-manager-789f6589d5-2b4sc\" (UID: \"f83accac-e306-4997-aba7-9276b9f9409c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.949254 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl9wr\" (UniqueName: \"kubernetes.io/projected/e0406bc0-0aff-4aa1-9f2a-e1af26b498b0-kube-api-access-kl9wr\") pod \"olm-operator-6b444d44fb-bkq5w\" (UID: \"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.959179 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2ftjk" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.980021 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h456\" (UniqueName: \"kubernetes.io/projected/13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b-kube-api-access-7h456\") pod \"ingress-canary-mn8dg\" (UID: \"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b\") " pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.982872 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7ce732b-7d7c-43f9-8944-72bd510a3be2-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mzkwr\" (UID: \"f7ce732b-7d7c-43f9-8944-72bd510a3be2\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.984667 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:47 crc kubenswrapper[4831]: I1124 08:17:47.984803 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kzxvh"] Nov 24 08:17:47 crc kubenswrapper[4831]: E1124 08:17:47.984994 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.484983124 +0000 UTC m=+142.360128277 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.016232 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dqwb\" (UniqueName: \"kubernetes.io/projected/ef536ac2-f27c-47f3-9571-b69f72d65178-kube-api-access-6dqwb\") pod \"service-ca-operator-777779d784-wn8vk\" (UID: \"ef536ac2-f27c-47f3-9571-b69f72d65178\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.023989 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxs8k\" (UniqueName: \"kubernetes.io/projected/1a934030-84e1-45c1-a58f-00b41a6d1b38-kube-api-access-qxs8k\") pod \"csi-hostpathplugin-7gmw8\" (UID: \"1a934030-84e1-45c1-a58f-00b41a6d1b38\") " pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.049631 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.073190 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9gml\" (UniqueName: \"kubernetes.io/projected/fcd58485-1abf-48ee-b3d7-50ed174205a3-kube-api-access-p9gml\") pod \"ingress-operator-5b745b69d9-w5rt4\" (UID: \"fcd58485-1abf-48ee-b3d7-50ed174205a3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.085394 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.085734 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.585709497 +0000 UTC m=+142.460854650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.100384 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.100781 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.128452 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.143619 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.157012 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.187117 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.187512 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.687497611 +0000 UTC m=+142.562642764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.187768 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.203040 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.238566 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mn8dg" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.278757 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.288214 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.289030 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.788960505 +0000 UTC m=+142.664105658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.361038 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.404012 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.404332 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:48.904296197 +0000 UTC m=+142.779441420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.403208 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fdjbw"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.499446 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.515113 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.515607 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.015586973 +0000 UTC m=+142.890732126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.515632 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4dbcf"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.617258 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.617638 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.117626124 +0000 UTC m=+142.992771267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.650766 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" event={"ID":"02514c19-c154-4055-ae24-eb2acaeb03d3","Type":"ContainerStarted","Data":"0002b6008954d0fafc2e3dc185e8cd4dab50f80c5fc9070ea24df8f85474afc6"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.669648 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" event={"ID":"984066f7-5187-41f4-96b8-9cacec55a5d3","Type":"ContainerStarted","Data":"e0dbc4ab556d07cb7489e38ba5549ee139d3f3d2e0ba45d5be3ef9676d0db2a4"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.686892 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" event={"ID":"df388cd7-d680-494b-b6f8-d1e6c51cd06b","Type":"ContainerStarted","Data":"faeef0790dfd33a15af69accaf86ffe6fc52e1a5fd21ef062a7225864150d399"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.687608 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vczfv"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.692862 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.698353 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rrssz"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.701636 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ft42l" event={"ID":"0cb983bd-e989-4487-b31a-c0328ca2960f","Type":"ContainerStarted","Data":"cba5d300fde15917f47cddc116b0e6e4ae4cbd9299ed6627b0edc5d51c5bb212"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.709100 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" event={"ID":"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed","Type":"ContainerStarted","Data":"94b75c9961125f3745bc52fb15309d9098f16ff1e3a56145740bae6202677558"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.718837 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.719178 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.21916057 +0000 UTC m=+143.094305723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.738258 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" event={"ID":"9e12a303-77ed-4145-b211-12926737b82e","Type":"ContainerStarted","Data":"cdeca20afd0aa18d3915332f0486c5a4d0084b22d6aabcef998e99553bd706c4"} Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.738385 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.744810 4831 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-k2k6z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.744868 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.746834 4831 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6lg2c container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.746888 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.746949 4831 patch_prober.go:28] interesting pod/console-operator-58897d9998-sq6n7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.746974 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" podUID="446d711b-f141-4823-8a50-b7f88ff704ea" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.747854 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw"] Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.822494 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.827223 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.327206283 +0000 UTC m=+143.202351636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.891942 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" podStartSLOduration=122.891920492 podStartE2EDuration="2m2.891920492s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:48.844790328 +0000 UTC m=+142.719935491" watchObservedRunningTime="2025-11-24 08:17:48.891920492 +0000 UTC m=+142.767065645" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.924274 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:48 crc kubenswrapper[4831]: E1124 08:17:48.925015 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.424995502 +0000 UTC m=+143.300140655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.935536 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9kqsc" podStartSLOduration=122.935514734 podStartE2EDuration="2m2.935514734s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:48.893309372 +0000 UTC m=+142.768454535" watchObservedRunningTime="2025-11-24 08:17:48.935514734 +0000 UTC m=+142.810659887" Nov 24 08:17:48 crc kubenswrapper[4831]: I1124 08:17:48.968608 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4szpt" podStartSLOduration=122.968573934 podStartE2EDuration="2m2.968573934s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:48.963251931 +0000 UTC m=+142.838397084" watchObservedRunningTime="2025-11-24 08:17:48.968573934 +0000 UTC m=+142.843719087" Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.027183 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.027537 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.527526307 +0000 UTC m=+143.402671450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: W1124 08:17:49.029145 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5db624a_81de_476b_8df0_24b1c2f0243f.slice/crio-a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224 WatchSource:0}: Error finding container a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224: Status 404 returned error can't find the container with id a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224 Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.044535 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5d6v2" podStartSLOduration=123.044509615 podStartE2EDuration="2m3.044509615s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:49.036880835 +0000 UTC m=+142.912025988" watchObservedRunningTime="2025-11-24 08:17:49.044509615 +0000 UTC m=+142.919654768" Nov 24 08:17:49 crc kubenswrapper[4831]: W1124 08:17:49.072636 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce29f16d_07cb_4d33_95bd_035d6d012ad2.slice/crio-66073475390a42067f4b6fba98d100b1acad572f9627ee6db3ff7b7a17126ae1 WatchSource:0}: Error finding container 66073475390a42067f4b6fba98d100b1acad572f9627ee6db3ff7b7a17126ae1: Status 404 returned error can't find the container with id 66073475390a42067f4b6fba98d100b1acad572f9627ee6db3ff7b7a17126ae1 Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.128016 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.128933 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.628906538 +0000 UTC m=+143.504051691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.129262 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.129637 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.629626419 +0000 UTC m=+143.504771572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.230742 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.231132 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.731118244 +0000 UTC m=+143.606263387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.332728 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.333013 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.83300224 +0000 UTC m=+143.708147393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.432192 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-prssr"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.434123 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.435709 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.935669179 +0000 UTC m=+143.810814322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.435786 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.448923 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:49.948898589 +0000 UTC m=+143.824043742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.538851 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.540825 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.039182022 +0000 UTC m=+143.914327175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.605502 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bq4m2"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.630669 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.640700 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.640984 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.140972175 +0000 UTC m=+144.016117318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.648741 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.654398 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-qkrmn" podStartSLOduration=123.65438092 podStartE2EDuration="2m3.65438092s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:49.654333229 +0000 UTC m=+143.529478402" watchObservedRunningTime="2025-11-24 08:17:49.65438092 +0000 UTC m=+143.529526073" Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.741387 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.742216 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.242200852 +0000 UTC m=+144.117346005 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.811628 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.851828 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.852149 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.35213357 +0000 UTC m=+144.227278713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.855990 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.871771 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" event={"ID":"df388cd7-d680-494b-b6f8-d1e6c51cd06b","Type":"ContainerStarted","Data":"2d667fe80d5e34fa0bb4455ab1193efdac94f99b41e2bfc7b4d7af58bffcced4"} Nov 24 08:17:49 crc kubenswrapper[4831]: I1124 08:17:49.953391 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:49 crc kubenswrapper[4831]: E1124 08:17:49.953818 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.45380256 +0000 UTC m=+144.328947713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.013257 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cx7sp" podStartSLOduration=124.013231396 podStartE2EDuration="2m4.013231396s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:49.996207067 +0000 UTC m=+143.871352220" watchObservedRunningTime="2025-11-24 08:17:50.013231396 +0000 UTC m=+143.888376549" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.055917 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.057029 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.557012903 +0000 UTC m=+144.432158056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.107448 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" event={"ID":"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8","Type":"ContainerStarted","Data":"31ec8ad315b2e674c5290f5fb5960d5de6a19528b10333d654a3193a571fe197"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.132931 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rrssz" event={"ID":"c6ffcfa0-86f7-41ca-934f-a633c2450a32","Type":"ContainerStarted","Data":"9bf70be45d462f0859b57739c3e36252875457e4827a3a4813175d0dda0e4933"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.160725 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.161166 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.661133923 +0000 UTC m=+144.536279076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.165449 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" event={"ID":"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9","Type":"ContainerStarted","Data":"2fcab9abd261f268194c4cfa98cfd8028dbcc3037b1592baf599378df21ae081"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.184892 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" podStartSLOduration=124.184873215 podStartE2EDuration="2m4.184873215s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:50.13072807 +0000 UTC m=+144.005873233" watchObservedRunningTime="2025-11-24 08:17:50.184873215 +0000 UTC m=+144.060018358" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.200115 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.220686 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" event={"ID":"ce29f16d-07cb-4d33-95bd-035d6d012ad2","Type":"ContainerStarted","Data":"66073475390a42067f4b6fba98d100b1acad572f9627ee6db3ff7b7a17126ae1"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.263960 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.264267 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.764254045 +0000 UTC m=+144.639399198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.265377 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" podStartSLOduration=124.265357827 podStartE2EDuration="2m4.265357827s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:50.262911467 +0000 UTC m=+144.138056720" watchObservedRunningTime="2025-11-24 08:17:50.265357827 +0000 UTC m=+144.140502970" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.291074 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ft42l" event={"ID":"0cb983bd-e989-4487-b31a-c0328ca2960f","Type":"ContainerStarted","Data":"2e9ddfeafa9be89937ff6869af73ef95c9554642849bce72913198eeb998bfc0"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.380357 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" event={"ID":"972d8e1c-6711-4b23-8014-8c463852b780","Type":"ContainerStarted","Data":"dd042dfe7c6aa07cef26eb48e3a70100d93dc6b5ee0da89dc165372e69d7c4b8"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.382018 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.383046 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.883030846 +0000 UTC m=+144.758175999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.450486 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" event={"ID":"a5db624a-81de-476b-8df0-24b1c2f0243f","Type":"ContainerStarted","Data":"a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.481786 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.484175 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.484481 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:50.9844679 +0000 UTC m=+144.859613053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.525439 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4bw9d" podStartSLOduration=124.482308028 podStartE2EDuration="2m4.482308028s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:50.400012714 +0000 UTC m=+144.275157877" watchObservedRunningTime="2025-11-24 08:17:50.482308028 +0000 UTC m=+144.357453181" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.564899 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.589616 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.589830 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pw597"] Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.590900 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.090873446 +0000 UTC m=+144.966018599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.663829 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.689364 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4dbcf" event={"ID":"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52","Type":"ContainerStarted","Data":"e7732e5e0d24ec6e124f629f3a7bf11ffbdc4185552e56c120fe81978c346526"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.689402 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4dbcf" event={"ID":"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52","Type":"ContainerStarted","Data":"22697affc317c59b9793a0ea40fd6a777860c0c9753683901e0b4815ac1aab1b"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.690453 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.690781 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.691178 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.191164706 +0000 UTC m=+145.066309859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.692030 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.692076 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.713378 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.718212 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.793560 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.795289 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.295272776 +0000 UTC m=+145.170417929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.838764 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" event={"ID":"dd19c608-b11b-4ef4-9efa-11710fec5a1a","Type":"ContainerStarted","Data":"bc2d3313a28d8e65442d73b614c6771eaf06fdacc6e9e401138e201fa982ae52"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.856235 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.884537 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.884586 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.899755 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:50 crc kubenswrapper[4831]: E1124 08:17:50.904575 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.404550845 +0000 UTC m=+145.279695998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:50 crc kubenswrapper[4831]: W1124 08:17:50.924448 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69b6963b_9605_46c1_b800_280539eec365.slice/crio-ec9cc9ef8761b4620c154e5c79c72f1506cc585096c9e62bfc31081903f0f56f WatchSource:0}: Error finding container ec9cc9ef8761b4620c154e5c79c72f1506cc585096c9e62bfc31081903f0f56f: Status 404 returned error can't find the container with id ec9cc9ef8761b4620c154e5c79c72f1506cc585096c9e62bfc31081903f0f56f Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.954183 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2ftjk" event={"ID":"e5984c81-33f3-47e5-9911-2c2b18f07f23","Type":"ContainerStarted","Data":"f345a053f6550e9b4cf0ccbab9971dc230aa585491cfdbce9a0f6b2ce3560e67"} Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.987293 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4"] Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.998004 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ft42l" podStartSLOduration=124.997983758 podStartE2EDuration="2m4.997983758s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:50.978843108 +0000 UTC m=+144.853988271" watchObservedRunningTime="2025-11-24 08:17:50.997983758 +0000 UTC m=+144.873128911" Nov 24 08:17:50 crc kubenswrapper[4831]: I1124 08:17:50.998805 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" event={"ID":"3aba0829-409c-4395-bf72-9eba0a55c877","Type":"ContainerStarted","Data":"0445a78c8b0983336e7e7b8e2a9b1f5f9e1aaa848ecac9e0d818a163f5f4e21e"} Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.008678 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" event={"ID":"fcc50e57-65ea-4a4a-aeb6-524b44ff9bed","Type":"ContainerStarted","Data":"3acc86a2645ebc9aeff6f2fd19304b603f1a552bcd9aca65b7966c6190ca7cf5"} Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.013084 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.013468 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.513424872 +0000 UTC m=+145.388570145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.085201 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sq6n7" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.114674 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.118662 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.618646214 +0000 UTC m=+145.493791367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.126558 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7gmw8"] Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.129424 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" podStartSLOduration=125.129412193 podStartE2EDuration="2m5.129412193s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:51.104849057 +0000 UTC m=+144.979994220" watchObservedRunningTime="2025-11-24 08:17:51.129412193 +0000 UTC m=+145.004557346" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.152901 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mn8dg"] Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.207144 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk"] Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.228378 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.228956 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.728935841 +0000 UTC m=+145.604080994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.237937 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" podStartSLOduration=125.237905169 podStartE2EDuration="2m5.237905169s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:51.208123733 +0000 UTC m=+145.083268886" watchObservedRunningTime="2025-11-24 08:17:51.237905169 +0000 UTC m=+145.113050322" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.333428 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.334292 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.834271856 +0000 UTC m=+145.709417009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.355026 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-4dbcf" podStartSLOduration=125.354987921 podStartE2EDuration="2m5.354987921s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:51.276271261 +0000 UTC m=+145.151416434" watchObservedRunningTime="2025-11-24 08:17:51.354987921 +0000 UTC m=+145.230133104" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.360845 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt"] Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.436746 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.438342 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:51.938285634 +0000 UTC m=+145.813430787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.541164 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.541521 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.041502748 +0000 UTC m=+145.916647901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.572573 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fdjbw" podStartSLOduration=125.57255573 podStartE2EDuration="2m5.57255573s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:51.36607046 +0000 UTC m=+145.241215633" watchObservedRunningTime="2025-11-24 08:17:51.57255573 +0000 UTC m=+145.447700883" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.573509 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2ftjk" podStartSLOduration=6.5735032570000005 podStartE2EDuration="6.573503257s" podCreationTimestamp="2025-11-24 08:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:51.568209325 +0000 UTC m=+145.443354488" watchObservedRunningTime="2025-11-24 08:17:51.573503257 +0000 UTC m=+145.448648410" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.645843 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.646420 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.146391981 +0000 UTC m=+146.021537134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.695461 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr"] Nov 24 08:17:51 crc kubenswrapper[4831]: W1124 08:17:51.740580 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7ce732b_7d7c_43f9_8944_72bd510a3be2.slice/crio-3a38a325fc5ea080f7a834d28c5bb0b50acf666e40d59a14ca6719d52fea6dd8 WatchSource:0}: Error finding container 3a38a325fc5ea080f7a834d28c5bb0b50acf666e40d59a14ca6719d52fea6dd8: Status 404 returned error can't find the container with id 3a38a325fc5ea080f7a834d28c5bb0b50acf666e40d59a14ca6719d52fea6dd8 Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.749352 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.749854 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.249834722 +0000 UTC m=+146.124979875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.772440 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-cc7tj" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.865442 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.866484 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.366455601 +0000 UTC m=+146.241600764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.867759 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:51 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:51 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:51 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.867820 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:51 crc kubenswrapper[4831]: I1124 08:17:51.973804 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:51 crc kubenswrapper[4831]: E1124 08:17:51.974075 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.474064002 +0000 UTC m=+146.349209155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.049733 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" event={"ID":"1a934030-84e1-45c1-a58f-00b41a6d1b38","Type":"ContainerStarted","Data":"75df56fae4aaa0c93790f881790c0ea659aa2146b472a949eb4d6f30b26dfc00"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.077775 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.078256 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.578237833 +0000 UTC m=+146.453382986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.100151 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" event={"ID":"34330ad7-9d56-4d26-bf45-15f34990ea8b","Type":"ContainerStarted","Data":"e641ac0547fc433a02627b5817f702d5afeae0ccebcb9d9e6c7a4469b3503134"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.100200 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" event={"ID":"34330ad7-9d56-4d26-bf45-15f34990ea8b","Type":"ContainerStarted","Data":"fed3dd305dc52d282bfce36a2a49b4a1b3fb2605d993d8f925101fc5d4236aca"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.120986 4831 generic.go:334] "Generic (PLEG): container finished" podID="02514c19-c154-4055-ae24-eb2acaeb03d3" containerID="da3bab4839a347939e9addcdefefa66c3b92a6f563e0cc377a1669ae43999de4" exitCode=0 Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.121093 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" event={"ID":"02514c19-c154-4055-ae24-eb2acaeb03d3","Type":"ContainerDied","Data":"da3bab4839a347939e9addcdefefa66c3b92a6f563e0cc377a1669ae43999de4"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.167157 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" event={"ID":"ff236b0b-a353-4f4e-9d87-06739909b3c2","Type":"ContainerStarted","Data":"59145cdabd2e4dfea08b2c82ff15534d7a0af711498acfbaeabc46106bfd1eeb"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.167215 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" event={"ID":"ff236b0b-a353-4f4e-9d87-06739909b3c2","Type":"ContainerStarted","Data":"32ffe203259cb325cfb641648d94d08649440a35bf94b2ed1de78df30f93bc13"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.168212 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.173821 4831 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5t9z8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.173932 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.181204 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.182662 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.682645492 +0000 UTC m=+146.557790635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.204065 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" event={"ID":"df388cd7-d680-494b-b6f8-d1e6c51cd06b","Type":"ContainerStarted","Data":"942ca37e69cacbade851916833b79002abd1ee8e069f0ed8e7da09054bf6e0c8"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.258693 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" event={"ID":"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0","Type":"ContainerStarted","Data":"0431dc33847fb4de2444f5653122d72e3968f75bfbfcb9670bc471977fbd5ef6"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.259782 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.285376 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.286611 4831 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bkq5w container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.286702 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" podUID="e0406bc0-0aff-4aa1-9f2a-e1af26b498b0" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.287167 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.787143853 +0000 UTC m=+146.662289006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.290563 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" podStartSLOduration=126.290546691 podStartE2EDuration="2m6.290546691s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.240179435 +0000 UTC m=+146.115324608" watchObservedRunningTime="2025-11-24 08:17:52.290546691 +0000 UTC m=+146.165691834" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.320289 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-z784l" podStartSLOduration=126.320255264 podStartE2EDuration="2m6.320255264s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.291034195 +0000 UTC m=+146.166179358" watchObservedRunningTime="2025-11-24 08:17:52.320255264 +0000 UTC m=+146.195400417" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.326593 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" event={"ID":"fba95cbe-ea61-442e-b35a-fc31277e4e7e","Type":"ContainerStarted","Data":"6f690d6dea2e2d4e7812f3be0f3b5772a836c84e573308e847d32f4aadd70df1"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.326648 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" event={"ID":"fba95cbe-ea61-442e-b35a-fc31277e4e7e","Type":"ContainerStarted","Data":"1aa1e9157ebb6a9228325c73ca4c00debea31d6c6118fc0870c9f6c724c66923"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.330400 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.334304 4831 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d8l7s container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.334416 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" podUID="fba95cbe-ea61-442e-b35a-fc31277e4e7e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.335778 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" podStartSLOduration=126.335754519 podStartE2EDuration="2m6.335754519s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.332017032 +0000 UTC m=+146.207162195" watchObservedRunningTime="2025-11-24 08:17:52.335754519 +0000 UTC m=+146.210899672" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.373583 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" podStartSLOduration=126.373563055 podStartE2EDuration="2m6.373563055s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.367714327 +0000 UTC m=+146.242859500" watchObservedRunningTime="2025-11-24 08:17:52.373563055 +0000 UTC m=+146.248708208" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.388072 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.394513 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.894483126 +0000 UTC m=+146.769628459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.402436 4831 generic.go:334] "Generic (PLEG): container finished" podID="984066f7-5187-41f4-96b8-9cacec55a5d3" containerID="6e20da108ee8463c1a90d01fa152d1c1b6dfbeeb8905332d89a3c711c371b798" exitCode=0 Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.402535 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" event={"ID":"984066f7-5187-41f4-96b8-9cacec55a5d3","Type":"ContainerDied","Data":"6e20da108ee8463c1a90d01fa152d1c1b6dfbeeb8905332d89a3c711c371b798"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.474804 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" event={"ID":"dd19c608-b11b-4ef4-9efa-11710fec5a1a","Type":"ContainerStarted","Data":"386fcb4a9f25dae8e70f72f2e5754fb4f9fdfc4a5781692f589804734216c89b"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.490398 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.492249 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:52.992208113 +0000 UTC m=+146.867353266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.568811 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-clzxl" podStartSLOduration=126.568781802 podStartE2EDuration="2m6.568781802s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.568736681 +0000 UTC m=+146.443881844" watchObservedRunningTime="2025-11-24 08:17:52.568781802 +0000 UTC m=+146.443926955" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.585869 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" event={"ID":"a5db624a-81de-476b-8df0-24b1c2f0243f","Type":"ContainerStarted","Data":"4f1974f587e8b3da79701df8c9721ad244c21bc0ed2d9c547747c05038f6ff4b"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.586404 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.592188 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.592526 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.092512724 +0000 UTC m=+146.967657877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.593688 4831 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4nv2q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" start-of-body= Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.593786 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": dial tcp 10.217.0.21:6443: connect: connection refused" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.688388 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mn8dg" event={"ID":"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b","Type":"ContainerStarted","Data":"e59b474f90c36761e601e988e441e07e58c9b306eb580691e51026fa32660f2a"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.698742 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.698940 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.19891537 +0000 UTC m=+147.074060523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.699069 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.701217 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.201204085 +0000 UTC m=+147.076349248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.706485 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" event={"ID":"8e68f5ca-dc2f-45a1-b5f5-e84fb92b6fa9","Type":"ContainerStarted","Data":"9e1cad56c09f283fe16b11ebef065261c4c42661d554c51c3d95fcc56e3eb53d"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.716991 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" podStartSLOduration=126.716976728 podStartE2EDuration="2m6.716976728s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.68916489 +0000 UTC m=+146.564310053" watchObservedRunningTime="2025-11-24 08:17:52.716976728 +0000 UTC m=+146.592121881" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.752660 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" event={"ID":"972d8e1c-6711-4b23-8014-8c463852b780","Type":"ContainerStarted","Data":"9c17254a2192b7c3d93c1a3a1ec937dde548533e5085142934b3a26db80ad565"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.789009 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vczfv" podStartSLOduration=125.788994367 podStartE2EDuration="2m5.788994367s" podCreationTimestamp="2025-11-24 08:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.7887779 +0000 UTC m=+146.663923053" watchObservedRunningTime="2025-11-24 08:17:52.788994367 +0000 UTC m=+146.664139520" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.789905 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-mn8dg" podStartSLOduration=7.789901153 podStartE2EDuration="7.789901153s" podCreationTimestamp="2025-11-24 08:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.753646242 +0000 UTC m=+146.628791395" watchObservedRunningTime="2025-11-24 08:17:52.789901153 +0000 UTC m=+146.665046306" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.800525 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.801839 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.301819505 +0000 UTC m=+147.176964658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.859531 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" event={"ID":"df09a1f6-dfe4-4fd2-94a0-280c6a701737","Type":"ContainerStarted","Data":"e360a72687789dbcbd5e4399206e256f87f16450704b1a152414d246d8340049"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.869892 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:52 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:52 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:52 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.869965 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.903088 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:52 crc kubenswrapper[4831]: E1124 08:17:52.903546 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.403528466 +0000 UTC m=+147.278673619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.955867 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" event={"ID":"69b6963b-9605-46c1-b800-280539eec365","Type":"ContainerStarted","Data":"ec9cc9ef8761b4620c154e5c79c72f1506cc585096c9e62bfc31081903f0f56f"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.962476 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" event={"ID":"3aba0829-409c-4395-bf72-9eba0a55c877","Type":"ContainerStarted","Data":"1677ede7029c48c569483560f969db8d2ce22cbd2bab39e5286fbbe7516aef19"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.962550 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" event={"ID":"3aba0829-409c-4395-bf72-9eba0a55c877","Type":"ContainerStarted","Data":"0ec43b291eefbd374a2a1a811fc1c6fbe390264b3800710feac9ea27a2446a1b"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.976419 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" event={"ID":"f1162ff1-0f44-4168-bcf6-945e5ef9adfd","Type":"ContainerStarted","Data":"3aaa65eef003cfa39b8ea85fc267d06de428754079aee823fbd852b96d62b6e5"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.976462 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" event={"ID":"f1162ff1-0f44-4168-bcf6-945e5ef9adfd","Type":"ContainerStarted","Data":"2a7439f9ac5b83ff6222316c3a6b09509829b5fb4210a377279bbb10d2d0496b"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.996530 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" event={"ID":"f83accac-e306-4997-aba7-9276b9f9409c","Type":"ContainerStarted","Data":"01f0749797f7a325245518f5eed9da246795033daa5a0f88dc46735538e75c5c"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.996581 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" event={"ID":"f83accac-e306-4997-aba7-9276b9f9409c","Type":"ContainerStarted","Data":"c995f4af553a80f8d39ed95df108926fdb859bba3bd7b17f3858d1ba44aef95d"} Nov 24 08:17:52 crc kubenswrapper[4831]: I1124 08:17:52.997147 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.006757 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-v29kz" podStartSLOduration=127.00673045 podStartE2EDuration="2m7.00673045s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:52.830032665 +0000 UTC m=+146.705177818" watchObservedRunningTime="2025-11-24 08:17:53.00673045 +0000 UTC m=+146.881875603" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.007973 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.008461 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.508435869 +0000 UTC m=+147.383581042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.023764 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rxwsw" event={"ID":"ce29f16d-07cb-4d33-95bd-035d6d012ad2","Type":"ContainerStarted","Data":"4eb37636c43dfc6c3e9f20d5acc20b18648bf33ad3d0bc1673f2f707b6f38891"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.050097 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" event={"ID":"fcd58485-1abf-48ee-b3d7-50ed174205a3","Type":"ContainerStarted","Data":"c4015e8b017d9672ab3a548ab2247635efa224380f344a164dd473b58b5525c7"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.052767 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" event={"ID":"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8","Type":"ContainerStarted","Data":"3d849344fc5e728db5d657089b41f189e43facce7c38200f34288d451b115c64"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.058569 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" event={"ID":"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82","Type":"ContainerStarted","Data":"b6552d5453205a9be043f8999b0c387ca6ba7a705ebeae72d8b3767a796ebab2"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.058595 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" event={"ID":"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82","Type":"ContainerStarted","Data":"56513a8a44bdbe05a610036d2c40f02e9ea2d113db7a8eafc0b76acf8f18c1c5"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.073046 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2ftjk" event={"ID":"e5984c81-33f3-47e5-9911-2c2b18f07f23","Type":"ContainerStarted","Data":"fab6a666d5d8c6070e2879cb12481fbf2e8c1b22397eb1bb6ab1fe084d3d8841"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.088282 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" podStartSLOduration=127.088249501 podStartE2EDuration="2m7.088249501s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.008044488 +0000 UTC m=+146.883189641" watchObservedRunningTime="2025-11-24 08:17:53.088249501 +0000 UTC m=+146.963394654" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.115227 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" event={"ID":"8d7c3f8f-edbc-449a-b891-3c91952dfa2f","Type":"ContainerStarted","Data":"acafd6246322ebfd2f972619f1c4a32b95d888f95a1acf562e9afb2bd759f808"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.116003 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.116502 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.616488602 +0000 UTC m=+147.491633755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.175824 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rrssz" event={"ID":"c6ffcfa0-86f7-41ca-934f-a633c2450a32","Type":"ContainerStarted","Data":"e01da7e0236daf6fcafbc68f84bbf61e163af4eba39c4f4120540058cc8b354e"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.176555 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rrssz" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.178884 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" podStartSLOduration=127.178854514 podStartE2EDuration="2m7.178854514s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.174017965 +0000 UTC m=+147.049163128" watchObservedRunningTime="2025-11-24 08:17:53.178854514 +0000 UTC m=+147.053999667" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.180654 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-prssr" podStartSLOduration=127.180643885 podStartE2EDuration="2m7.180643885s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.104614541 +0000 UTC m=+146.979759694" watchObservedRunningTime="2025-11-24 08:17:53.180643885 +0000 UTC m=+147.055789038" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.184273 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" event={"ID":"ef536ac2-f27c-47f3-9571-b69f72d65178","Type":"ContainerStarted","Data":"83d3bd465783ed72b33dafe4aa61e78693f246460850ecf356d62e672d86d5d2"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.220022 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.220149 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.720124919 +0000 UTC m=+147.595270062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.220339 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.221015 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.720992894 +0000 UTC m=+147.596138047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.239827 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" event={"ID":"d4f0d807-8634-4c21-a4b8-924ad3463b0c","Type":"ContainerStarted","Data":"c6d55356ee0a1aa2d6ad3f358d7265cee08195b772bf5bc2c0264187f011a249"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.241115 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.243031 4831 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-jkrqt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.243085 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" podUID="d4f0d807-8634-4c21-a4b8-924ad3463b0c" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.258882 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.258947 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.259624 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" event={"ID":"f7ce732b-7d7c-43f9-8944-72bd510a3be2","Type":"ContainerStarted","Data":"3a38a325fc5ea080f7a834d28c5bb0b50acf666e40d59a14ca6719d52fea6dd8"} Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.295632 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nrcst" podStartSLOduration=127.295589896 podStartE2EDuration="2m7.295589896s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.219729418 +0000 UTC m=+147.094874591" watchObservedRunningTime="2025-11-24 08:17:53.295589896 +0000 UTC m=+147.170735049" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.321189 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.323089 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.823066775 +0000 UTC m=+147.698211928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.345323 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" podStartSLOduration=127.345284363 podStartE2EDuration="2m7.345284363s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.342782682 +0000 UTC m=+147.217927845" watchObservedRunningTime="2025-11-24 08:17:53.345284363 +0000 UTC m=+147.220429516" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.355515 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rrssz" podStartSLOduration=8.355489187 podStartE2EDuration="8.355489187s" podCreationTimestamp="2025-11-24 08:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.29677677 +0000 UTC m=+147.171921933" watchObservedRunningTime="2025-11-24 08:17:53.355489187 +0000 UTC m=+147.230634340" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.423469 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.423813 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:53.923794848 +0000 UTC m=+147.798940001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.429705 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" podStartSLOduration=127.429688328 podStartE2EDuration="2m7.429688328s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.42874403 +0000 UTC m=+147.303889193" watchObservedRunningTime="2025-11-24 08:17:53.429688328 +0000 UTC m=+147.304833491" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.430279 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" podStartSLOduration=127.430272254 podStartE2EDuration="2m7.430272254s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:53.397020769 +0000 UTC m=+147.272165942" watchObservedRunningTime="2025-11-24 08:17:53.430272254 +0000 UTC m=+147.305417407" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.524548 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.524894 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.024877922 +0000 UTC m=+147.900023075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.626069 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.626465 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.126449878 +0000 UTC m=+148.001595031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.727655 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.728013 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.227997304 +0000 UTC m=+148.103142457 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.832134 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.832744 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.332716232 +0000 UTC m=+148.207861585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.867588 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:53 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:53 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:53 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.867733 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.933889 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.934091 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.434064143 +0000 UTC m=+148.309209296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:53 crc kubenswrapper[4831]: I1124 08:17:53.934212 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:53 crc kubenswrapper[4831]: E1124 08:17:53.934610 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.434598418 +0000 UTC m=+148.309743571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.035877 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.036103 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.536068902 +0000 UTC m=+148.411214065 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.036426 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.036756 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.536741582 +0000 UTC m=+148.411886815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.137780 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.137976 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.637952738 +0000 UTC m=+148.513097891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.138040 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.138509 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.638497014 +0000 UTC m=+148.513642167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.241260 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.241644 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.741629856 +0000 UTC m=+148.616775009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.286807 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" event={"ID":"1a934030-84e1-45c1-a58f-00b41a6d1b38","Type":"ContainerStarted","Data":"d74cde4f65eb61ba9e1de62dba401bb261e680c6da96415e3ab0b20d7c29d6d4"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.294648 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" event={"ID":"f7ce732b-7d7c-43f9-8944-72bd510a3be2","Type":"ContainerStarted","Data":"9998e573406de576c1504ea6494bc8bbf389c22564dd8ab978e5401f5badf95f"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.297787 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-546qh" event={"ID":"69b6963b-9605-46c1-b800-280539eec365","Type":"ContainerStarted","Data":"24d21b28e13b715551e6338158b0ad4021d4ff9e09aaeb6d6489de5ac30a2f43"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.312683 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" event={"ID":"fcd58485-1abf-48ee-b3d7-50ed174205a3","Type":"ContainerStarted","Data":"245ce5f953ca1e639ae1787d1833e70b11449fdeccfc3e7bbaefe60cfc565865"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.312767 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" event={"ID":"fcd58485-1abf-48ee-b3d7-50ed174205a3","Type":"ContainerStarted","Data":"6ae01a06e5dc8548fc5bd8ebfbb244d6b13fb9d6b29f036364a411aa75f11d44"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.322797 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" event={"ID":"3fab9f8d-f413-444f-b32b-d3f09cc3b2b8","Type":"ContainerStarted","Data":"393d8c0861f7aaa9c4b8e578731fef57bc22b5e78fa9e5191d4da8e4d3438320"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.342640 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.347285 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.8472702 +0000 UTC m=+148.722415353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.351659 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" event={"ID":"8d7c3f8f-edbc-449a-b891-3c91952dfa2f","Type":"ContainerStarted","Data":"ebb91b4dd95ec32293a55e838d9fcba19fdce23a01e7a250543a43eb5860c72b"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.351719 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" event={"ID":"8d7c3f8f-edbc-449a-b891-3c91952dfa2f","Type":"ContainerStarted","Data":"fcd2ea4de6d08563ed6e7fdb057e9bd4cd35e78aedfc1a4101b55e0fd22ec362"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.361176 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" event={"ID":"984066f7-5187-41f4-96b8-9cacec55a5d3","Type":"ContainerStarted","Data":"86328fb2c00ab95e7149ed59dc23378d5835bdb9f82eea922055a5f05a48b904"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.369829 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wn8vk" event={"ID":"ef536ac2-f27c-47f3-9571-b69f72d65178","Type":"ContainerStarted","Data":"c64f20dcec269b5118adda244396e043e21a11b056adddec7008d41e5e425a1f"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.374008 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mn8dg" event={"ID":"13ee520d-ccf4-4899-a3c1-d8cd5cb8f64b","Type":"ContainerStarted","Data":"4cc59c3e977c86e9acbc75b36c16871119bc4b6f6761e6c8ae052283adfce4f5"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.391291 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" event={"ID":"f83accac-e306-4997-aba7-9276b9f9409c","Type":"ContainerStarted","Data":"f528da9a9a1ac68bc8046ad494e34418108dd4bb38f5b844a41bc3352b2b4101"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.415167 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" event={"ID":"df09a1f6-dfe4-4fd2-94a0-280c6a701737","Type":"ContainerStarted","Data":"cf082f7172b1ad4b48dc522b3e86993cfec11cad7ff262dac59d1f3a349d6e4c"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.415234 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" event={"ID":"df09a1f6-dfe4-4fd2-94a0-280c6a701737","Type":"ContainerStarted","Data":"457d906f9b6c7c51b72fd3d81ddd4192bc6b271de4757f9ba86970fffe864782"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.431215 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" event={"ID":"e0406bc0-0aff-4aa1-9f2a-e1af26b498b0","Type":"ContainerStarted","Data":"5b2b685a7b8c22e586f8d6546e6bda83a74fba37fe390a081e455749d145799a"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.432556 4831 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bkq5w container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.432688 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" podUID="e0406bc0-0aff-4aa1-9f2a-e1af26b498b0" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.439625 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" event={"ID":"d4f0d807-8634-4c21-a4b8-924ad3463b0c","Type":"ContainerStarted","Data":"72767f8325e469e9f1ba89f6215731804237d907f2816fbb62f4694f316a19c6"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.442954 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" event={"ID":"34330ad7-9d56-4d26-bf45-15f34990ea8b","Type":"ContainerStarted","Data":"08fdb5085945b19ec748047a2ec228b1e3c1fc510fbf1838e1295d870c5c6347"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.444702 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.445091 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.945059149 +0000 UTC m=+148.820204302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.445558 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.445847 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" event={"ID":"02514c19-c154-4055-ae24-eb2acaeb03d3","Type":"ContainerStarted","Data":"b936313b36e84261cab63ed6dcd9c83f07da238f9b87f3c092f31fff11c76343"} Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.447500 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:54.947490288 +0000 UTC m=+148.822635441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.460392 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mzkwr" podStartSLOduration=128.460373318 podStartE2EDuration="2m8.460373318s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.34832661 +0000 UTC m=+148.223471773" watchObservedRunningTime="2025-11-24 08:17:54.460373318 +0000 UTC m=+148.335518461" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.461575 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bq4m2" podStartSLOduration=128.461569533 podStartE2EDuration="2m8.461569533s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.459084181 +0000 UTC m=+148.334229344" watchObservedRunningTime="2025-11-24 08:17:54.461569533 +0000 UTC m=+148.336714686" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.474505 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rrssz" event={"ID":"c6ffcfa0-86f7-41ca-934f-a633c2450a32","Type":"ContainerStarted","Data":"9d8718e92051d2734adf929c8da2a29245fb1a1b63b592e4fa76c7e479e5accf"} Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.481274 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.481351 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.481384 4831 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5t9z8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.481471 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.546232 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.546415 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.046386809 +0000 UTC m=+148.921531962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.549709 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.556762 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.056738666 +0000 UTC m=+148.931883819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.640241 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-w5rt4" podStartSLOduration=128.640224894 podStartE2EDuration="2m8.640224894s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.577673437 +0000 UTC m=+148.452818590" watchObservedRunningTime="2025-11-24 08:17:54.640224894 +0000 UTC m=+148.515370047" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.650733 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.651056 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.151040775 +0000 UTC m=+149.026185928 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.694146 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pw597" podStartSLOduration=128.694129402 podStartE2EDuration="2m8.694129402s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.642675204 +0000 UTC m=+148.517820367" watchObservedRunningTime="2025-11-24 08:17:54.694129402 +0000 UTC m=+148.569274565" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.694974 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" podStartSLOduration=128.694967896 podStartE2EDuration="2m8.694967896s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.692640119 +0000 UTC m=+148.567785282" watchObservedRunningTime="2025-11-24 08:17:54.694967896 +0000 UTC m=+148.570113059" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.736278 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zxwd2" podStartSLOduration=128.736263822 podStartE2EDuration="2m8.736263822s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.73583753 +0000 UTC m=+148.610982713" watchObservedRunningTime="2025-11-24 08:17:54.736263822 +0000 UTC m=+148.611408965" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.752984 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.753421 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.753455 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.753496 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.754038 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.254021732 +0000 UTC m=+149.129166885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.755573 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.777901 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.795944 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.820639 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-jkrqt" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.838228 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hwbrt" podStartSLOduration=128.8382133 podStartE2EDuration="2m8.8382133s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.83472006 +0000 UTC m=+148.709865213" watchObservedRunningTime="2025-11-24 08:17:54.8382133 +0000 UTC m=+148.713358453" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.838658 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" podStartSLOduration=128.838652513 podStartE2EDuration="2m8.838652513s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:54.808062994 +0000 UTC m=+148.683208147" watchObservedRunningTime="2025-11-24 08:17:54.838652513 +0000 UTC m=+148.713797666" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.854307 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.854631 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.855265 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.355227169 +0000 UTC m=+149.230372332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.870656 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:54 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:54 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:54 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.870715 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.871634 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.909279 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.920932 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.926503 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:54 crc kubenswrapper[4831]: I1124 08:17:54.963477 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:54 crc kubenswrapper[4831]: E1124 08:17:54.963767 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.463756746 +0000 UTC m=+149.338901889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.066499 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.066699 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.566674342 +0000 UTC m=+149.441819495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.066944 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.067243 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.567236758 +0000 UTC m=+149.442381911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.169456 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.169832 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.669817664 +0000 UTC m=+149.544962807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.272881 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.273214 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.773202753 +0000 UTC m=+149.648347896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.374504 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.375065 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.875050168 +0000 UTC m=+149.750195321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.475679 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.475733 4831 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-d8l7s container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.475801 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" podUID="fba95cbe-ea61-442e-b35a-fc31277e4e7e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.476038 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:55.976024898 +0000 UTC m=+149.851170051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.480467 4831 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4nv2q container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.480581 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.21:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.486759 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" event={"ID":"984066f7-5187-41f4-96b8-9cacec55a5d3","Type":"ContainerStarted","Data":"98a0e6618934b2e070728056b63f4207b517eee7ff87dd686cbf1aeac27468e7"} Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.490607 4831 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5t9z8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.490947 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.501459 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bkq5w" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.579776 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.581263 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.08124715 +0000 UTC m=+149.956392303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.681791 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.682151 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.182137858 +0000 UTC m=+150.057283011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.782997 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.797837 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.29781159 +0000 UTC m=+150.172956743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.805533 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.860220 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:55 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:55 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:55 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.860269 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.885971 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.886255 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.38624339 +0000 UTC m=+150.261388543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.987626 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-d8l7s" Nov 24 08:17:55 crc kubenswrapper[4831]: I1124 08:17:55.989905 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:55 crc kubenswrapper[4831]: E1124 08:17:55.990260 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.490242307 +0000 UTC m=+150.365387460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.091115 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.091452 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.591440753 +0000 UTC m=+150.466585896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.193299 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.193639 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.693624498 +0000 UTC m=+150.568769641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.276194 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.296389 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.296677 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.796663897 +0000 UTC m=+150.671809050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.397141 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.398125 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:56.898110451 +0000 UTC m=+150.773255604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.448959 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.449845 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.470384 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.503799 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrtm2\" (UniqueName: \"kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.503835 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.503871 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.503897 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.504151 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.004140876 +0000 UTC m=+150.879286019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.521289 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.543214 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" event={"ID":"1a934030-84e1-45c1-a58f-00b41a6d1b38","Type":"ContainerStarted","Data":"2fe50f4cb1687587b8ee82c91e770ba26327022d5361060334e528d63ee7ca2c"} Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.606530 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.606745 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.106729073 +0000 UTC m=+150.981874226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.607031 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrtm2\" (UniqueName: \"kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.607084 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.607107 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.607416 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.609050 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.609467 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.610434 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.110190262 +0000 UTC m=+150.985335415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.635981 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.636026 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.636467 4831 patch_prober.go:28] interesting pod/console-f9d7485db-9kqsc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.636524 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9kqsc" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.638138 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.642673 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.643709 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: W1124 08:17:56.676157 4831 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.676197 4831 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.677861 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.711103 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.711415 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgxph\" (UniqueName: \"kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.711449 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.711516 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.712235 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.212218472 +0000 UTC m=+151.087363625 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.722356 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrtm2\" (UniqueName: \"kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2\") pod \"community-operators-fcscl\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.808859 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.814982 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.815061 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.815093 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgxph\" (UniqueName: \"kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.815116 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.815518 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.815724 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.815942 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.315931731 +0000 UTC m=+151.191076884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.865008 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:56 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:56 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:56 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.865275 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.874539 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.876578 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.878386 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.983641 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.984217 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.984438 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.984566 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmxc7\" (UniqueName: \"kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:56 crc kubenswrapper[4831]: I1124 08:17:56.984597 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:56 crc kubenswrapper[4831]: E1124 08:17:56.992864 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.492842942 +0000 UTC m=+151.367988095 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.008757 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgxph\" (UniqueName: \"kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph\") pod \"certified-operators-kfvtn\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.068286 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.069203 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.097290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.097360 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.097662 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.597649762 +0000 UTC m=+151.472794915 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.097665 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmxc7\" (UniqueName: \"kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.097702 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.098135 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.098201 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.119596 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.179130 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmxc7\" (UniqueName: \"kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7\") pod \"community-operators-6rwmr\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.202075 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.202363 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shwwj\" (UniqueName: \"kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.202417 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.202439 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.202969 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.702949225 +0000 UTC m=+151.578094378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: W1124 08:17:57.223943 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-0fd719479e8b08dbc8e28b948b01e7a10465141819f8b88532b3545f45f3e671 WatchSource:0}: Error finding container 0fd719479e8b08dbc8e28b948b01e7a10465141819f8b88532b3545f45f3e671: Status 404 returned error can't find the container with id 0fd719479e8b08dbc8e28b948b01e7a10465141819f8b88532b3545f45f3e671 Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.308133 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shwwj\" (UniqueName: \"kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.308180 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.308198 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.308222 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.308473 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.808462346 +0000 UTC m=+151.683607499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.311788 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.314409 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.367987 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.380355 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shwwj\" (UniqueName: \"kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj\") pod \"certified-operators-87tx9\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.408727 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.409127 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:57.909107436 +0000 UTC m=+151.784252589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.472675 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.473387 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.482873 4831 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kzxvh container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.482916 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" podUID="984066f7-5187-41f4-96b8-9cacec55a5d3" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.7:8443/livez\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.485701 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.486386 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.502575 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.511240 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.511590 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.011578679 +0000 UTC m=+151.886723832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.584510 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.584570 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.584961 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.584981 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.609694 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0738dc9980de7c8bd25cfab9b6f60e0863bd1d3ca4068e8b839aaf5fe1bff8b3"} Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.609731 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0fd719479e8b08dbc8e28b948b01e7a10465141819f8b88532b3545f45f3e671"} Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.614084 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.614713 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.114697231 +0000 UTC m=+151.989842384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.647656 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a8137648b039a88e26e1355ae6c5c0dac26aceefedc2351b5a017e976039be08"} Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.691534 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d03524d1fd9f138cc2af08354497d3efa580da6b61303fb2f85d408d7cd072e3"} Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.691599 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0747128787ffd58d446638b381b31cda6222f59cac52cf913c6c16e4d5276d58"} Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.702528 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cdqhk" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.717335 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.717686 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.217675348 +0000 UTC m=+152.092820501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.819994 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.821206 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.321190841 +0000 UTC m=+152.196335994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.860080 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.861902 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:57 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:57 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:57 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.861950 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.909045 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.911633 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.914072 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.916532 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:17:57 crc kubenswrapper[4831]: I1124 08:17:57.933595 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:57 crc kubenswrapper[4831]: E1124 08:17:57.934916 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.434900247 +0000 UTC m=+152.310045400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.036310 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.036582 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.536540066 +0000 UTC m=+152.411685219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.036747 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.037118 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.537110533 +0000 UTC m=+152.412255686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.139117 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.139492 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.639477963 +0000 UTC m=+152.514623116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.242146 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.242494 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.742482431 +0000 UTC m=+152.617627584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.276843 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.365797 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.366383 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.866369449 +0000 UTC m=+152.741514602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.408190 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.408246 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.454307 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.467408 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.467715 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:58.96770378 +0000 UTC m=+152.842848933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.568234 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.568597 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.068582027 +0000 UTC m=+152.943727170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.670355 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.671204 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.171190794 +0000 UTC m=+153.046335957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.703543 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerStarted","Data":"60c51541fe30c7131ca6424132983f24a50291e93f09116e04e5bc1f7231cb3b"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.711252 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ac9a02a34742a09b3f62f09c6603562989313934771315f42808b1e48527acfd"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.711764 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.713450 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerStarted","Data":"c8516e15e9036b7e7880e11a38ed87a300a6f64a4f6c410ae6c6a24757024e49"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.713495 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerStarted","Data":"eb5ef42dbdc73aa0d6e037edce02184d9b76a582b34ea07be386a9bc5d2adc00"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.719057 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.731108 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" event={"ID":"1a934030-84e1-45c1-a58f-00b41a6d1b38","Type":"ContainerStarted","Data":"1cd852449ed85fc9fd4eb68d9c7e759f4f8b2d99d6c91cd8350f8e530f09bd65"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.731146 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" event={"ID":"1a934030-84e1-45c1-a58f-00b41a6d1b38","Type":"ContainerStarted","Data":"17eb868e9ff414731b121ab9e3426dda07e7c3b2fc301d8ba9e114353c173a12"} Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.754741 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.770812 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7gmw8" podStartSLOduration=13.770796845 podStartE2EDuration="13.770796845s" podCreationTimestamp="2025-11-24 08:17:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:17:58.769220969 +0000 UTC m=+152.644366142" watchObservedRunningTime="2025-11-24 08:17:58.770796845 +0000 UTC m=+152.645941998" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.771898 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.772175 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.272163784 +0000 UTC m=+153.147308937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.782193 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.783328 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.785597 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.804617 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.860641 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:58 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:58 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:58 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.860695 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.874308 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.875340 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.375309956 +0000 UTC m=+153.250455109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.991280 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.991530 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5scr\" (UniqueName: \"kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.991583 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:58 crc kubenswrapper[4831]: I1124 08:17:58.991612 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:58 crc kubenswrapper[4831]: E1124 08:17:58.991726 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.491712539 +0000 UTC m=+153.366857692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.073574 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.094135 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5scr\" (UniqueName: \"kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.094191 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.094223 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.094253 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.094664 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.095273 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.595257293 +0000 UTC m=+153.470402446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.095673 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.124340 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5scr\" (UniqueName: \"kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr\") pod \"redhat-marketplace-tks88\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.184725 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.186266 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.199371 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.201399 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.701364961 +0000 UTC m=+153.576510114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.209448 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.303354 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.303680 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.303809 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zhvv\" (UniqueName: \"kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.303910 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.304310 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.804298327 +0000 UTC m=+153.679443480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.356126 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.356765 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.363199 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.363338 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.375892 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.404408 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.404628 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.904608028 +0000 UTC m=+153.779753181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.404839 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zhvv\" (UniqueName: \"kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.404953 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.405063 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.405183 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.405270 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.405361 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.405809 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:17:59.905794862 +0000 UTC m=+153.780940015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.406056 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.406464 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.420115 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.431201 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zhvv\" (UniqueName: \"kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv\") pod \"redhat-marketplace-jh829\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.465583 4831 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.506058 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.506195 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.006176895 +0000 UTC m=+153.881322048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.506343 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.506369 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.506422 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.506429 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.506959 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.006941247 +0000 UTC m=+153.882086460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.547626 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.548715 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.593423 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.594856 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.598900 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.610458 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.610665 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.610758 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.110745708 +0000 UTC m=+153.985890851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.610966 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.611035 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.611067 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkn4\" (UniqueName: \"kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.611084 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.611374 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.111365426 +0000 UTC m=+153.986510579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.711924 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.712138 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.712262 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkn4\" (UniqueName: \"kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.712566 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.212541142 +0000 UTC m=+154.087686295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.712674 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.712836 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.713338 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.731731 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.742872 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkn4\" (UniqueName: \"kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4\") pod \"redhat-operators-4lkcs\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.783207 4831 generic.go:334] "Generic (PLEG): container finished" podID="184380c6-d131-4447-9549-f93708be879c" containerID="c8516e15e9036b7e7880e11a38ed87a300a6f64a4f6c410ae6c6a24757024e49" exitCode=0 Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.783289 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerDied","Data":"c8516e15e9036b7e7880e11a38ed87a300a6f64a4f6c410ae6c6a24757024e49"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.787640 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.788837 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.796496 4831 generic.go:334] "Generic (PLEG): container finished" podID="a174e67d-8201-4967-88d2-2c47b6976df3" containerID="5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c" exitCode=0 Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.796578 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerDied","Data":"5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.796603 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerStarted","Data":"4847fe85585e20d38263014777e5581c9a4347b6e8e51f9a5ff4e4e596048e45"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.797370 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.808117 4831 generic.go:334] "Generic (PLEG): container finished" podID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerID="7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e" exitCode=0 Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.808170 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerDied","Data":"7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.808194 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerStarted","Data":"d2aea5b209dba840b1b306027e35b59a358895f739405758c7d8ad32116cf625"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.815035 4831 generic.go:334] "Generic (PLEG): container finished" podID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerID="a87c4870317c13890501ababbbf62113271795cfa10eba59d382fc13290902a6" exitCode=0 Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.815518 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerDied","Data":"a87c4870317c13890501ababbbf62113271795cfa10eba59d382fc13290902a6"} Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.815943 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.816093 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.816230 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq25j\" (UniqueName: \"kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.816292 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.816644 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.316628191 +0000 UTC m=+154.191773344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.877416 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:17:59 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:17:59 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:17:59 crc kubenswrapper[4831]: healthz check failed Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.877484 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.921266 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.921450 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.421422281 +0000 UTC m=+154.296567434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.921605 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq25j\" (UniqueName: \"kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.921686 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.921728 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.921842 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.923176 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: E1124 08:17:59.923196 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.423179632 +0000 UTC m=+154.298324875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.923368 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.958191 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq25j\" (UniqueName: \"kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j\") pod \"redhat-operators-kd4td\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:17:59 crc kubenswrapper[4831]: I1124 08:17:59.969309 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.029871 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:18:00 crc kubenswrapper[4831]: E1124 08:18:00.030231 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.530181655 +0000 UTC m=+154.405326818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.109875 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.120576 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.123413 4831 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-24T08:17:59.465817756Z","Handler":null,"Name":""} Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.131993 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:00 crc kubenswrapper[4831]: E1124 08:18:00.132332 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-24 08:18:00.632305178 +0000 UTC m=+154.507450331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-6l4dx" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 24 08:18:00 crc kubenswrapper[4831]: W1124 08:18:00.155811 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b8f6f3_d0d2_45ef_ae1e_f73b6f7e6f8f.slice/crio-76bc74ab6490b03f1e678a0018d2bcdd1b1f8b963985d3bb68e92ff2b89dc6a0 WatchSource:0}: Error finding container 76bc74ab6490b03f1e678a0018d2bcdd1b1f8b963985d3bb68e92ff2b89dc6a0: Status 404 returned error can't find the container with id 76bc74ab6490b03f1e678a0018d2bcdd1b1f8b963985d3bb68e92ff2b89dc6a0 Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.183383 4831 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.183451 4831 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.232962 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.251095 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.340240 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.359258 4831 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.359297 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.451949 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.742545 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.743506 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-6l4dx\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.826669 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.858058 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:00 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:00 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:00 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.858098 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.884244 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerStarted","Data":"2c88a55fe586d5bdc75add55371be7b9896a00f2eb4c3e2b3130d6642ef1653d"} Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.887834 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerStarted","Data":"368574d392f7b562a56034c927a6d7826abb0468a6224a92c2102e76d4316bb4"} Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.890684 4831 generic.go:334] "Generic (PLEG): container finished" podID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerID="ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790" exitCode=0 Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.890723 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerDied","Data":"ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790"} Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.890746 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerStarted","Data":"76bc74ab6490b03f1e678a0018d2bcdd1b1f8b963985d3bb68e92ff2b89dc6a0"} Nov 24 08:18:00 crc kubenswrapper[4831]: I1124 08:18:00.925048 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.073398 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.074547 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.469540 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.695153 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.695947 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.697547 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.699281 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.699524 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.778888 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.778959 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.857831 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:01 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:01 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:01 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.857881 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.880361 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.880448 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.880580 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.902902 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"41a66876-e225-4faa-9215-9d00c029f7c7","Type":"ContainerStarted","Data":"23a3426e51c94f1b3a66c4b584ead0f4ae5fe968f70be1e171919e4535e09181"} Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.902950 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"41a66876-e225-4faa-9215-9d00c029f7c7","Type":"ContainerStarted","Data":"4b8e97707816eb781f9adffeab6da1efe8ab5243ef343dbe8a5e12f6092d5968"} Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.906306 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.927223 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.927209547 podStartE2EDuration="2.927209547s" podCreationTimestamp="2025-11-24 08:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:18:01.914581134 +0000 UTC m=+155.789726297" watchObservedRunningTime="2025-11-24 08:18:01.927209547 +0000 UTC m=+155.802354700" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.928030 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" event={"ID":"75023939-0385-4a3f-b734-fad2cb5fdc34","Type":"ContainerStarted","Data":"77d8907db75eb830886edf44e19dbb3bfd71262f3c9d716d050398c15c6bc647"} Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.928069 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" event={"ID":"75023939-0385-4a3f-b734-fad2cb5fdc34","Type":"ContainerStarted","Data":"2f57bb767a01824ae625fabc3b07d5590f059d43341cacae4f9092af86bec74d"} Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.928231 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.965907 4831 generic.go:334] "Generic (PLEG): container finished" podID="414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" containerID="b6552d5453205a9be043f8999b0c387ca6ba7a705ebeae72d8b3767a796ebab2" exitCode=0 Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.966022 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" event={"ID":"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82","Type":"ContainerDied","Data":"b6552d5453205a9be043f8999b0c387ca6ba7a705ebeae72d8b3767a796ebab2"} Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.995580 4831 generic.go:334] "Generic (PLEG): container finished" podID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerID="1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c" exitCode=0 Nov 24 08:18:01 crc kubenswrapper[4831]: I1124 08:18:01.995677 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerDied","Data":"1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c"} Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.001060 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" podStartSLOduration=136.001040628 podStartE2EDuration="2m16.001040628s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:18:01.965846467 +0000 UTC m=+155.840991630" watchObservedRunningTime="2025-11-24 08:18:02.001040628 +0000 UTC m=+155.876185781" Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.012443 4831 generic.go:334] "Generic (PLEG): container finished" podID="9c997027-92eb-4878-beb2-6fbb939737ce" containerID="8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c" exitCode=0 Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.012503 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerDied","Data":"8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c"} Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.012529 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerStarted","Data":"64911dce70f50faae47cd7c75af6b9f01727d4b466973ba3dbb9017d2b19d638"} Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.027116 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.051672 4831 generic.go:334] "Generic (PLEG): container finished" podID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerID="c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c" exitCode=0 Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.051731 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerDied","Data":"c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c"} Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.479023 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.483969 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kzxvh" Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.660666 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rrssz" Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.862195 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:02 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:02 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:02 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:02 crc kubenswrapper[4831]: I1124 08:18:02.862247 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.024609 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.111914 4831 generic.go:334] "Generic (PLEG): container finished" podID="41a66876-e225-4faa-9215-9d00c029f7c7" containerID="23a3426e51c94f1b3a66c4b584ead0f4ae5fe968f70be1e171919e4535e09181" exitCode=0 Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.112039 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"41a66876-e225-4faa-9215-9d00c029f7c7","Type":"ContainerDied","Data":"23a3426e51c94f1b3a66c4b584ead0f4ae5fe968f70be1e171919e4535e09181"} Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.137238 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2","Type":"ContainerStarted","Data":"96ab5425970abf325f8faf9f5ece9f902d042cbd686c00a537810d1951bf523e"} Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.662122 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.827664 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume\") pod \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.827768 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume\") pod \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.827825 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7q7q\" (UniqueName: \"kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q\") pod \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\" (UID: \"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82\") " Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.840056 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume" (OuterVolumeSpecName: "config-volume") pod "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" (UID: "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.857635 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q" (OuterVolumeSpecName: "kube-api-access-p7q7q") pod "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" (UID: "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82"). InnerVolumeSpecName "kube-api-access-p7q7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.858929 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:03 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:03 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:03 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.858987 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.859181 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" (UID: "414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.932959 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7q7q\" (UniqueName: \"kubernetes.io/projected/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-kube-api-access-p7q7q\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.932998 4831 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:03 crc kubenswrapper[4831]: I1124 08:18:03.933007 4831 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.266367 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.273741 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw" event={"ID":"414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82","Type":"ContainerDied","Data":"56513a8a44bdbe05a610036d2c40f02e9ea2d113db7a8eafc0b76acf8f18c1c5"} Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.273785 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56513a8a44bdbe05a610036d2c40f02e9ea2d113db7a8eafc0b76acf8f18c1c5" Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.841485 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.880871 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:04 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:04 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:04 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.880932 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.982540 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access\") pod \"41a66876-e225-4faa-9215-9d00c029f7c7\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.982707 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir\") pod \"41a66876-e225-4faa-9215-9d00c029f7c7\" (UID: \"41a66876-e225-4faa-9215-9d00c029f7c7\") " Nov 24 08:18:04 crc kubenswrapper[4831]: I1124 08:18:04.983028 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "41a66876-e225-4faa-9215-9d00c029f7c7" (UID: "41a66876-e225-4faa-9215-9d00c029f7c7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.007548 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "41a66876-e225-4faa-9215-9d00c029f7c7" (UID: "41a66876-e225-4faa-9215-9d00c029f7c7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.084995 4831 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/41a66876-e225-4faa-9215-9d00c029f7c7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.085048 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/41a66876-e225-4faa-9215-9d00c029f7c7-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.321088 4831 generic.go:334] "Generic (PLEG): container finished" podID="9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" containerID="be4024550dcb3861cd7d19bc3c3e8029b7dc7fa1af17257121f6ed4d4f4fdeea" exitCode=0 Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.321189 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2","Type":"ContainerDied","Data":"be4024550dcb3861cd7d19bc3c3e8029b7dc7fa1af17257121f6ed4d4f4fdeea"} Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.327274 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"41a66876-e225-4faa-9215-9d00c029f7c7","Type":"ContainerDied","Data":"4b8e97707816eb781f9adffeab6da1efe8ab5243ef343dbe8a5e12f6092d5968"} Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.327340 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b8e97707816eb781f9adffeab6da1efe8ab5243ef343dbe8a5e12f6092d5968" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.327398 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.858078 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:05 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:05 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:05 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:05 crc kubenswrapper[4831]: I1124 08:18:05.858138 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:06 crc kubenswrapper[4831]: I1124 08:18:06.633650 4831 patch_prober.go:28] interesting pod/console-f9d7485db-9kqsc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Nov 24 08:18:06 crc kubenswrapper[4831]: I1124 08:18:06.633991 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9kqsc" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Nov 24 08:18:06 crc kubenswrapper[4831]: I1124 08:18:06.856185 4831 patch_prober.go:28] interesting pod/router-default-5444994796-ft42l container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 24 08:18:06 crc kubenswrapper[4831]: [-]has-synced failed: reason withheld Nov 24 08:18:06 crc kubenswrapper[4831]: [+]process-running ok Nov 24 08:18:06 crc kubenswrapper[4831]: healthz check failed Nov 24 08:18:06 crc kubenswrapper[4831]: I1124 08:18:06.856272 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ft42l" podUID="0cb983bd-e989-4487-b31a-c0328ca2960f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 24 08:18:06 crc kubenswrapper[4831]: I1124 08:18:06.911498 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.018709 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir\") pod \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.018915 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" (UID: "9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.018931 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access\") pod \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\" (UID: \"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2\") " Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.019420 4831 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.046622 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" (UID: "9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.120744 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.383804 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2","Type":"ContainerDied","Data":"96ab5425970abf325f8faf9f5ece9f902d042cbd686c00a537810d1951bf523e"} Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.383878 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96ab5425970abf325f8faf9f5ece9f902d042cbd686c00a537810d1951bf523e" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.383972 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.585515 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.585606 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.586119 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.586049 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.862358 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:18:07 crc kubenswrapper[4831]: I1124 08:18:07.868932 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ft42l" Nov 24 08:18:08 crc kubenswrapper[4831]: I1124 08:18:08.648425 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:18:08 crc kubenswrapper[4831]: I1124 08:18:08.675077 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/650855b7-20e6-4261-a212-4bc75a671b96-metrics-certs\") pod \"network-metrics-daemon-k45nx\" (UID: \"650855b7-20e6-4261-a212-4bc75a671b96\") " pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:18:08 crc kubenswrapper[4831]: I1124 08:18:08.809386 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-k45nx" Nov 24 08:18:09 crc kubenswrapper[4831]: I1124 08:18:09.817030 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-k45nx"] Nov 24 08:18:10 crc kubenswrapper[4831]: I1124 08:18:10.483985 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k45nx" event={"ID":"650855b7-20e6-4261-a212-4bc75a671b96","Type":"ContainerStarted","Data":"5b82c278eb853a47fbf358296b81f2a213894ab11316a78a0863cfe51380d41e"} Nov 24 08:18:11 crc kubenswrapper[4831]: I1124 08:18:11.499562 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k45nx" event={"ID":"650855b7-20e6-4261-a212-4bc75a671b96","Type":"ContainerStarted","Data":"26e7571eb2a7590bb48ce7806f7cd0630092fd66b86001fe36ecaf25a58f30c0"} Nov 24 08:18:12 crc kubenswrapper[4831]: I1124 08:18:12.533753 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-k45nx" event={"ID":"650855b7-20e6-4261-a212-4bc75a671b96","Type":"ContainerStarted","Data":"89bddf66ab2782f9453b9d407294f786584d6002cfb0a4a4040fe8796692d244"} Nov 24 08:18:12 crc kubenswrapper[4831]: I1124 08:18:12.580780 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-k45nx" podStartSLOduration=146.580755398 podStartE2EDuration="2m26.580755398s" podCreationTimestamp="2025-11-24 08:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:18:12.578840653 +0000 UTC m=+166.453985826" watchObservedRunningTime="2025-11-24 08:18:12.580755398 +0000 UTC m=+166.455900551" Nov 24 08:18:16 crc kubenswrapper[4831]: I1124 08:18:16.638140 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:18:16 crc kubenswrapper[4831]: I1124 08:18:16.645946 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.584133 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.584210 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.586511 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.586638 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.586693 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.587108 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"e7732e5e0d24ec6e124f629f3a7bf11ffbdc4185552e56c120fe81978c346526"} pod="openshift-console/downloads-7954f5f757-4dbcf" containerMessage="Container download-server failed liveness probe, will be restarted" Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.587219 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" containerID="cri-o://e7732e5e0d24ec6e124f629f3a7bf11ffbdc4185552e56c120fe81978c346526" gracePeriod=2 Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.587357 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:17 crc kubenswrapper[4831]: I1124 08:18:17.587381 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:18 crc kubenswrapper[4831]: I1124 08:18:18.603659 4831 generic.go:334] "Generic (PLEG): container finished" podID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerID="e7732e5e0d24ec6e124f629f3a7bf11ffbdc4185552e56c120fe81978c346526" exitCode=0 Nov 24 08:18:18 crc kubenswrapper[4831]: I1124 08:18:18.603723 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4dbcf" event={"ID":"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52","Type":"ContainerDied","Data":"e7732e5e0d24ec6e124f629f3a7bf11ffbdc4185552e56c120fe81978c346526"} Nov 24 08:18:20 crc kubenswrapper[4831]: I1124 08:18:20.832819 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:18:27 crc kubenswrapper[4831]: I1124 08:18:27.585014 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:27 crc kubenswrapper[4831]: I1124 08:18:27.586102 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:28 crc kubenswrapper[4831]: I1124 08:18:28.208865 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2b4sc" Nov 24 08:18:28 crc kubenswrapper[4831]: I1124 08:18:28.401467 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:18:28 crc kubenswrapper[4831]: I1124 08:18:28.401590 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:18:34 crc kubenswrapper[4831]: I1124 08:18:34.942586 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 24 08:18:37 crc kubenswrapper[4831]: I1124 08:18:37.584578 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:37 crc kubenswrapper[4831]: I1124 08:18:37.585105 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.077461 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.077891 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bq25j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kd4td_openshift-marketplace(7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.079563 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kd4td" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.120108 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.120288 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-shwwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-87tx9_openshift-marketplace(4d2b0ef8-641f-462a-be15-8ecb8f06db6a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.121606 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-87tx9" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.127764 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.127867 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nrtm2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-fcscl_openshift-marketplace(184380c6-d131-4447-9549-f93708be879c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.129457 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-fcscl" podUID="184380c6-d131-4447-9549-f93708be879c" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.212233 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.212680 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9qkn4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4lkcs_openshift-marketplace(9c997027-92eb-4878-beb2-6fbb939737ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.216429 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4lkcs" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.743891 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4dbcf" event={"ID":"4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52","Type":"ContainerStarted","Data":"4fb1c87ab70227fcd95a4f19d9cdc91b5d68ba16becb3083a8dad24d105f90a8"} Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.744079 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.744622 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.744657 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.746590 4831 generic.go:334] "Generic (PLEG): container finished" podID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerID="18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4" exitCode=0 Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.746667 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerDied","Data":"18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4"} Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.750259 4831 generic.go:334] "Generic (PLEG): container finished" podID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerID="eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33" exitCode=0 Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.750377 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerDied","Data":"eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33"} Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.758788 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerStarted","Data":"c73b0b0a057c66e3f0417b4e770e2f08948253872948aa75443580dd74ce6a4d"} Nov 24 08:18:39 crc kubenswrapper[4831]: I1124 08:18:39.766454 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerStarted","Data":"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce"} Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.768101 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4lkcs" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.768437 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kd4td" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.771095 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-87tx9" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" Nov 24 08:18:39 crc kubenswrapper[4831]: E1124 08:18:39.773319 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-fcscl" podUID="184380c6-d131-4447-9549-f93708be879c" Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.769803 4831 generic.go:334] "Generic (PLEG): container finished" podID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerID="c73b0b0a057c66e3f0417b4e770e2f08948253872948aa75443580dd74ce6a4d" exitCode=0 Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.769866 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerDied","Data":"c73b0b0a057c66e3f0417b4e770e2f08948253872948aa75443580dd74ce6a4d"} Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.774884 4831 generic.go:334] "Generic (PLEG): container finished" podID="a174e67d-8201-4967-88d2-2c47b6976df3" containerID="702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce" exitCode=0 Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.774992 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerDied","Data":"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce"} Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.775855 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:40 crc kubenswrapper[4831]: I1124 08:18:40.775891 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:44 crc kubenswrapper[4831]: I1124 08:18:44.793781 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerStarted","Data":"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb"} Nov 24 08:18:45 crc kubenswrapper[4831]: I1124 08:18:45.802286 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerStarted","Data":"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1"} Nov 24 08:18:46 crc kubenswrapper[4831]: I1124 08:18:46.809717 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerStarted","Data":"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb"} Nov 24 08:18:46 crc kubenswrapper[4831]: I1124 08:18:46.830503 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tks88" podStartSLOduration=4.994624584 podStartE2EDuration="48.830483086s" podCreationTimestamp="2025-11-24 08:17:58 +0000 UTC" firstStartedPulling="2025-11-24 08:18:00.892364826 +0000 UTC m=+154.767509989" lastFinishedPulling="2025-11-24 08:18:44.728223338 +0000 UTC m=+198.603368491" observedRunningTime="2025-11-24 08:18:46.828473898 +0000 UTC m=+200.703619081" watchObservedRunningTime="2025-11-24 08:18:46.830483086 +0000 UTC m=+200.705628239" Nov 24 08:18:46 crc kubenswrapper[4831]: I1124 08:18:46.831120 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jh829" podStartSLOduration=5.672186544 podStartE2EDuration="47.831114314s" podCreationTimestamp="2025-11-24 08:17:59 +0000 UTC" firstStartedPulling="2025-11-24 08:18:02.083019822 +0000 UTC m=+155.958164965" lastFinishedPulling="2025-11-24 08:18:44.241947582 +0000 UTC m=+198.117092735" observedRunningTime="2025-11-24 08:18:44.810619285 +0000 UTC m=+198.685764448" watchObservedRunningTime="2025-11-24 08:18:46.831114314 +0000 UTC m=+200.706259467" Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.585302 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.585408 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.585460 4831 patch_prober.go:28] interesting pod/downloads-7954f5f757-4dbcf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.585542 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4dbcf" podUID="4e15f7d6-b0c3-4ded-b0b4-3f1ddfacfc52" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.842270 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfvtn" podStartSLOduration=5.194931745 podStartE2EDuration="51.842246943s" podCreationTimestamp="2025-11-24 08:17:56 +0000 UTC" firstStartedPulling="2025-11-24 08:17:59.806564422 +0000 UTC m=+153.681709575" lastFinishedPulling="2025-11-24 08:18:46.45387961 +0000 UTC m=+200.329024773" observedRunningTime="2025-11-24 08:18:47.841341197 +0000 UTC m=+201.716486370" watchObservedRunningTime="2025-11-24 08:18:47.842246943 +0000 UTC m=+201.717392096" Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.917268 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:18:47 crc kubenswrapper[4831]: I1124 08:18:47.917374 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:18:48 crc kubenswrapper[4831]: I1124 08:18:48.827180 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerStarted","Data":"0122e27fea26d1ce4866f18927226f788b3c21c2907edf54b328c0091804b2bc"} Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.407040 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.407214 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.495799 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.498153 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kfvtn" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="registry-server" probeResult="failure" output=< Nov 24 08:18:49 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:18:49 crc kubenswrapper[4831]: > Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.548737 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.548793 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.615707 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.857416 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6rwmr" podStartSLOduration=6.221702886 podStartE2EDuration="53.857397479s" podCreationTimestamp="2025-11-24 08:17:56 +0000 UTC" firstStartedPulling="2025-11-24 08:17:59.818218777 +0000 UTC m=+153.693363930" lastFinishedPulling="2025-11-24 08:18:47.45391337 +0000 UTC m=+201.329058523" observedRunningTime="2025-11-24 08:18:49.855359431 +0000 UTC m=+203.730504604" watchObservedRunningTime="2025-11-24 08:18:49.857397479 +0000 UTC m=+203.732542632" Nov 24 08:18:49 crc kubenswrapper[4831]: I1124 08:18:49.892704 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:50 crc kubenswrapper[4831]: I1124 08:18:50.887476 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.211655 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.212750 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jh829" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="registry-server" containerID="cri-o://0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb" gracePeriod=2 Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.733762 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.857839 4831 generic.go:334] "Generic (PLEG): container finished" podID="9c997027-92eb-4878-beb2-6fbb939737ce" containerID="51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2" exitCode=0 Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.857912 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerDied","Data":"51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2"} Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.869591 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerStarted","Data":"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0"} Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.872217 4831 generic.go:334] "Generic (PLEG): container finished" podID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerID="0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb" exitCode=0 Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.872248 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerDied","Data":"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb"} Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.872265 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jh829" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.872279 4831 scope.go:117] "RemoveContainer" containerID="0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.872269 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jh829" event={"ID":"9f460c3a-18c9-46d2-b590-f86729f598a5","Type":"ContainerDied","Data":"368574d392f7b562a56034c927a6d7826abb0468a6224a92c2102e76d4316bb4"} Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.889419 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities\") pod \"9f460c3a-18c9-46d2-b590-f86729f598a5\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.889592 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zhvv\" (UniqueName: \"kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv\") pod \"9f460c3a-18c9-46d2-b590-f86729f598a5\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.889626 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content\") pod \"9f460c3a-18c9-46d2-b590-f86729f598a5\" (UID: \"9f460c3a-18c9-46d2-b590-f86729f598a5\") " Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.891657 4831 scope.go:117] "RemoveContainer" containerID="18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.892608 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities" (OuterVolumeSpecName: "utilities") pod "9f460c3a-18c9-46d2-b590-f86729f598a5" (UID: "9f460c3a-18c9-46d2-b590-f86729f598a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.901380 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv" (OuterVolumeSpecName: "kube-api-access-4zhvv") pod "9f460c3a-18c9-46d2-b590-f86729f598a5" (UID: "9f460c3a-18c9-46d2-b590-f86729f598a5"). InnerVolumeSpecName "kube-api-access-4zhvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.927437 4831 scope.go:117] "RemoveContainer" containerID="c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.928249 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f460c3a-18c9-46d2-b590-f86729f598a5" (UID: "9f460c3a-18c9-46d2-b590-f86729f598a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.982449 4831 scope.go:117] "RemoveContainer" containerID="0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb" Nov 24 08:18:52 crc kubenswrapper[4831]: E1124 08:18:52.983545 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb\": container with ID starting with 0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb not found: ID does not exist" containerID="0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.983590 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb"} err="failed to get container status \"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb\": rpc error: code = NotFound desc = could not find container \"0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb\": container with ID starting with 0f273784da7f62e5d7cec7a084999e1953c23f5a91ef6587135b58940bb640cb not found: ID does not exist" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.983621 4831 scope.go:117] "RemoveContainer" containerID="18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4" Nov 24 08:18:52 crc kubenswrapper[4831]: E1124 08:18:52.984052 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4\": container with ID starting with 18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4 not found: ID does not exist" containerID="18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.984206 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4"} err="failed to get container status \"18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4\": rpc error: code = NotFound desc = could not find container \"18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4\": container with ID starting with 18227a64599c212f4b7f87fc1deb709656854ce9b0f4a6726fb4dc36760e64f4 not found: ID does not exist" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.984381 4831 scope.go:117] "RemoveContainer" containerID="c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c" Nov 24 08:18:52 crc kubenswrapper[4831]: E1124 08:18:52.986704 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c\": container with ID starting with c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c not found: ID does not exist" containerID="c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.986810 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c"} err="failed to get container status \"c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c\": rpc error: code = NotFound desc = could not find container \"c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c\": container with ID starting with c7bf858aad179b260990d3e8786487517c57f7d36fdab6c4c9908433fdf9377c not found: ID does not exist" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.991780 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zhvv\" (UniqueName: \"kubernetes.io/projected/9f460c3a-18c9-46d2-b590-f86729f598a5-kube-api-access-4zhvv\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.992066 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:52 crc kubenswrapper[4831]: I1124 08:18:52.992147 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f460c3a-18c9-46d2-b590-f86729f598a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.192062 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.197296 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jh829"] Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.878348 4831 generic.go:334] "Generic (PLEG): container finished" podID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerID="9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0" exitCode=0 Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.878415 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerDied","Data":"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0"} Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.883028 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerStarted","Data":"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a"} Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.889428 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerStarted","Data":"154792157e9769e87de2c3400c07b30b925fffbf778e45716ac29f24c1cf143f"} Nov 24 08:18:53 crc kubenswrapper[4831]: I1124 08:18:53.919582 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4lkcs" podStartSLOduration=3.354367325 podStartE2EDuration="54.919567005s" podCreationTimestamp="2025-11-24 08:17:59 +0000 UTC" firstStartedPulling="2025-11-24 08:18:02.030047681 +0000 UTC m=+155.905192834" lastFinishedPulling="2025-11-24 08:18:53.595247361 +0000 UTC m=+207.470392514" observedRunningTime="2025-11-24 08:18:53.916782135 +0000 UTC m=+207.791927308" watchObservedRunningTime="2025-11-24 08:18:53.919567005 +0000 UTC m=+207.794712148" Nov 24 08:18:54 crc kubenswrapper[4831]: I1124 08:18:54.899911 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" path="/var/lib/kubelet/pods/9f460c3a-18c9-46d2-b590-f86729f598a5/volumes" Nov 24 08:18:54 crc kubenswrapper[4831]: I1124 08:18:54.906076 4831 generic.go:334] "Generic (PLEG): container finished" podID="184380c6-d131-4447-9549-f93708be879c" containerID="154792157e9769e87de2c3400c07b30b925fffbf778e45716ac29f24c1cf143f" exitCode=0 Nov 24 08:18:54 crc kubenswrapper[4831]: I1124 08:18:54.907854 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerDied","Data":"154792157e9769e87de2c3400c07b30b925fffbf778e45716ac29f24c1cf143f"} Nov 24 08:18:54 crc kubenswrapper[4831]: I1124 08:18:54.923585 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerStarted","Data":"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95"} Nov 24 08:18:55 crc kubenswrapper[4831]: I1124 08:18:55.930031 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerStarted","Data":"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a"} Nov 24 08:18:55 crc kubenswrapper[4831]: I1124 08:18:55.953539 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-87tx9" podStartSLOduration=5.282025318 podStartE2EDuration="59.95352204s" podCreationTimestamp="2025-11-24 08:17:56 +0000 UTC" firstStartedPulling="2025-11-24 08:17:59.818593738 +0000 UTC m=+153.693738891" lastFinishedPulling="2025-11-24 08:18:54.49009047 +0000 UTC m=+208.365235613" observedRunningTime="2025-11-24 08:18:55.952166162 +0000 UTC m=+209.827311315" watchObservedRunningTime="2025-11-24 08:18:55.95352204 +0000 UTC m=+209.828667193" Nov 24 08:18:56 crc kubenswrapper[4831]: I1124 08:18:56.936116 4831 generic.go:334] "Generic (PLEG): container finished" podID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerID="11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95" exitCode=0 Nov 24 08:18:56 crc kubenswrapper[4831]: I1124 08:18:56.936159 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerDied","Data":"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95"} Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.313266 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.313582 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.362891 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.596946 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-4dbcf" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.915053 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.915092 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.942707 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerStarted","Data":"7080ffc57f89eed628c3030ab52a09d2966b6c64e98dc17a84add7d601ec1f50"} Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.969609 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fcscl" podStartSLOduration=3.695620726 podStartE2EDuration="1m1.969590343s" podCreationTimestamp="2025-11-24 08:17:56 +0000 UTC" firstStartedPulling="2025-11-24 08:17:58.718697658 +0000 UTC m=+152.593842811" lastFinishedPulling="2025-11-24 08:18:56.992667275 +0000 UTC m=+210.867812428" observedRunningTime="2025-11-24 08:18:57.967868763 +0000 UTC m=+211.843013916" watchObservedRunningTime="2025-11-24 08:18:57.969590343 +0000 UTC m=+211.844735506" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.979100 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:18:57 crc kubenswrapper[4831]: I1124 08:18:57.992892 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.005936 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.053508 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.401828 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.402138 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.402290 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.403095 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.403251 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860" gracePeriod=600 Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.949652 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860" exitCode=0 Nov 24 08:18:58 crc kubenswrapper[4831]: I1124 08:18:58.949736 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860"} Nov 24 08:18:59 crc kubenswrapper[4831]: I1124 08:18:59.957410 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerStarted","Data":"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6"} Nov 24 08:18:59 crc kubenswrapper[4831]: I1124 08:18:59.959877 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac"} Nov 24 08:18:59 crc kubenswrapper[4831]: I1124 08:18:59.970950 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:18:59 crc kubenswrapper[4831]: I1124 08:18:59.971352 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:18:59 crc kubenswrapper[4831]: I1124 08:18:59.985206 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kd4td" podStartSLOduration=3.5446453 podStartE2EDuration="1m0.98516826s" podCreationTimestamp="2025-11-24 08:17:59 +0000 UTC" firstStartedPulling="2025-11-24 08:18:02.000569984 +0000 UTC m=+155.875715137" lastFinishedPulling="2025-11-24 08:18:59.441092944 +0000 UTC m=+213.316238097" observedRunningTime="2025-11-24 08:18:59.98098813 +0000 UTC m=+213.856133283" watchObservedRunningTime="2025-11-24 08:18:59.98516826 +0000 UTC m=+213.860313413" Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.012496 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.012774 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6rwmr" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="registry-server" containerID="cri-o://0122e27fea26d1ce4866f18927226f788b3c21c2907edf54b328c0091804b2bc" gracePeriod=2 Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.121399 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.121499 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.968356 4831 generic.go:334] "Generic (PLEG): container finished" podID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerID="0122e27fea26d1ce4866f18927226f788b3c21c2907edf54b328c0091804b2bc" exitCode=0 Nov 24 08:19:00 crc kubenswrapper[4831]: I1124 08:19:00.968683 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerDied","Data":"0122e27fea26d1ce4866f18927226f788b3c21c2907edf54b328c0091804b2bc"} Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.026275 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.044851 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4lkcs" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="registry-server" probeResult="failure" output=< Nov 24 08:19:01 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:19:01 crc kubenswrapper[4831]: > Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.167448 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kd4td" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="registry-server" probeResult="failure" output=< Nov 24 08:19:01 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:19:01 crc kubenswrapper[4831]: > Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.200227 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities\") pod \"350d3b8d-c680-4373-87a8-e1c22d302c64\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.200282 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmxc7\" (UniqueName: \"kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7\") pod \"350d3b8d-c680-4373-87a8-e1c22d302c64\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.200434 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content\") pod \"350d3b8d-c680-4373-87a8-e1c22d302c64\" (UID: \"350d3b8d-c680-4373-87a8-e1c22d302c64\") " Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.202620 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities" (OuterVolumeSpecName: "utilities") pod "350d3b8d-c680-4373-87a8-e1c22d302c64" (UID: "350d3b8d-c680-4373-87a8-e1c22d302c64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.211518 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7" (OuterVolumeSpecName: "kube-api-access-bmxc7") pod "350d3b8d-c680-4373-87a8-e1c22d302c64" (UID: "350d3b8d-c680-4373-87a8-e1c22d302c64"). InnerVolumeSpecName "kube-api-access-bmxc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.269040 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "350d3b8d-c680-4373-87a8-e1c22d302c64" (UID: "350d3b8d-c680-4373-87a8-e1c22d302c64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.301986 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.302018 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/350d3b8d-c680-4373-87a8-e1c22d302c64-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.302032 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmxc7\" (UniqueName: \"kubernetes.io/projected/350d3b8d-c680-4373-87a8-e1c22d302c64-kube-api-access-bmxc7\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.976670 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6rwmr" event={"ID":"350d3b8d-c680-4373-87a8-e1c22d302c64","Type":"ContainerDied","Data":"60c51541fe30c7131ca6424132983f24a50291e93f09116e04e5bc1f7231cb3b"} Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.977014 4831 scope.go:117] "RemoveContainer" containerID="0122e27fea26d1ce4866f18927226f788b3c21c2907edf54b328c0091804b2bc" Nov 24 08:19:01 crc kubenswrapper[4831]: I1124 08:19:01.977149 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6rwmr" Nov 24 08:19:02 crc kubenswrapper[4831]: I1124 08:19:02.001408 4831 scope.go:117] "RemoveContainer" containerID="c73b0b0a057c66e3f0417b4e770e2f08948253872948aa75443580dd74ce6a4d" Nov 24 08:19:02 crc kubenswrapper[4831]: I1124 08:19:02.013289 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:19:02 crc kubenswrapper[4831]: I1124 08:19:02.022114 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6rwmr"] Nov 24 08:19:02 crc kubenswrapper[4831]: I1124 08:19:02.029253 4831 scope.go:117] "RemoveContainer" containerID="a87c4870317c13890501ababbbf62113271795cfa10eba59d382fc13290902a6" Nov 24 08:19:02 crc kubenswrapper[4831]: I1124 08:19:02.898684 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" path="/var/lib/kubelet/pods/350d3b8d-c680-4373-87a8-e1c22d302c64/volumes" Nov 24 08:19:06 crc kubenswrapper[4831]: I1124 08:19:06.809973 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:06 crc kubenswrapper[4831]: I1124 08:19:06.810924 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:06 crc kubenswrapper[4831]: I1124 08:19:06.859845 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:07 crc kubenswrapper[4831]: I1124 08:19:07.047100 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:07 crc kubenswrapper[4831]: I1124 08:19:07.969196 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:19:08 crc kubenswrapper[4831]: I1124 08:19:08.411923 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:19:08 crc kubenswrapper[4831]: I1124 08:19:08.412557 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-87tx9" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="registry-server" containerID="cri-o://49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a" gracePeriod=2 Nov 24 08:19:08 crc kubenswrapper[4831]: I1124 08:19:08.860426 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.012399 4831 generic.go:334] "Generic (PLEG): container finished" podID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerID="49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a" exitCode=0 Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.012595 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerDied","Data":"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a"} Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.012811 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-87tx9" event={"ID":"4d2b0ef8-641f-462a-be15-8ecb8f06db6a","Type":"ContainerDied","Data":"d2aea5b209dba840b1b306027e35b59a358895f739405758c7d8ad32116cf625"} Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.012833 4831 scope.go:117] "RemoveContainer" containerID="49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.012692 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-87tx9" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.019995 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shwwj\" (UniqueName: \"kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj\") pod \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.020101 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities\") pod \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.020131 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content\") pod \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\" (UID: \"4d2b0ef8-641f-462a-be15-8ecb8f06db6a\") " Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.021642 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities" (OuterVolumeSpecName: "utilities") pod "4d2b0ef8-641f-462a-be15-8ecb8f06db6a" (UID: "4d2b0ef8-641f-462a-be15-8ecb8f06db6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.033306 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj" (OuterVolumeSpecName: "kube-api-access-shwwj") pod "4d2b0ef8-641f-462a-be15-8ecb8f06db6a" (UID: "4d2b0ef8-641f-462a-be15-8ecb8f06db6a"). InnerVolumeSpecName "kube-api-access-shwwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.035784 4831 scope.go:117] "RemoveContainer" containerID="9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.074455 4831 scope.go:117] "RemoveContainer" containerID="7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.095935 4831 scope.go:117] "RemoveContainer" containerID="49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a" Nov 24 08:19:09 crc kubenswrapper[4831]: E1124 08:19:09.096712 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a\": container with ID starting with 49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a not found: ID does not exist" containerID="49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.096766 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a"} err="failed to get container status \"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a\": rpc error: code = NotFound desc = could not find container \"49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a\": container with ID starting with 49078df3e068c629db1d46c6e8ffe10aee5b1d3623463789ab4733ac4909238a not found: ID does not exist" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.096797 4831 scope.go:117] "RemoveContainer" containerID="9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0" Nov 24 08:19:09 crc kubenswrapper[4831]: E1124 08:19:09.097283 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0\": container with ID starting with 9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0 not found: ID does not exist" containerID="9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.097366 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0"} err="failed to get container status \"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0\": rpc error: code = NotFound desc = could not find container \"9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0\": container with ID starting with 9059765eac9814c9f5e2aee81879c28fa7e9fc7f6ebb3eb120ba20ff4e3458b0 not found: ID does not exist" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.097384 4831 scope.go:117] "RemoveContainer" containerID="7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e" Nov 24 08:19:09 crc kubenswrapper[4831]: E1124 08:19:09.097902 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e\": container with ID starting with 7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e not found: ID does not exist" containerID="7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.097946 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e"} err="failed to get container status \"7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e\": rpc error: code = NotFound desc = could not find container \"7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e\": container with ID starting with 7536e5ac82532792655f8e8324b7163d2a0a4c9d3577a9c26add2db7d52a265e not found: ID does not exist" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.101471 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d2b0ef8-641f-462a-be15-8ecb8f06db6a" (UID: "4d2b0ef8-641f-462a-be15-8ecb8f06db6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.121961 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.121992 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.122009 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shwwj\" (UniqueName: \"kubernetes.io/projected/4d2b0ef8-641f-462a-be15-8ecb8f06db6a-kube-api-access-shwwj\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.343678 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:19:09 crc kubenswrapper[4831]: I1124 08:19:09.348505 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-87tx9"] Nov 24 08:19:10 crc kubenswrapper[4831]: I1124 08:19:10.022752 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:19:10 crc kubenswrapper[4831]: I1124 08:19:10.084560 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:19:10 crc kubenswrapper[4831]: I1124 08:19:10.170752 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:10 crc kubenswrapper[4831]: I1124 08:19:10.234593 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:10 crc kubenswrapper[4831]: I1124 08:19:10.901645 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" path="/var/lib/kubelet/pods/4d2b0ef8-641f-462a-be15-8ecb8f06db6a/volumes" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.083536 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.084639 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfvtn" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="registry-server" containerID="cri-o://3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.098721 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.099106 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fcscl" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="registry-server" containerID="cri-o://7080ffc57f89eed628c3030ab52a09d2966b6c64e98dc17a84add7d601ec1f50" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.110576 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.110981 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" containerID="cri-o://59145cdabd2e4dfea08b2c82ff15534d7a0af711498acfbaeabc46106bfd1eeb" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.121081 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.121379 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tks88" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="registry-server" containerID="cri-o://60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.134482 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.135175 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4lkcs" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="registry-server" containerID="cri-o://347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.151378 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.151825 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kd4td" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="registry-server" containerID="cri-o://683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6" gracePeriod=30 Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.155268 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fcz4s"] Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.156978 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157000 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157013 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157020 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157039 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157045 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157054 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157060 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157073 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157080 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157088 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157094 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157103 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41a66876-e225-4faa-9215-9d00c029f7c7" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157111 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="41a66876-e225-4faa-9215-9d00c029f7c7" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157123 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157132 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157144 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157150 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157178 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157185 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="extract-utilities" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157195 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157203 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="extract-content" Nov 24 08:19:12 crc kubenswrapper[4831]: E1124 08:19:12.157210 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" containerName="collect-profiles" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157217 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" containerName="collect-profiles" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157335 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="41a66876-e225-4faa-9215-9d00c029f7c7" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157345 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8d74cf-de78-4af6-9e54-2ebc6eed9ca2" containerName="pruner" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157360 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" containerName="collect-profiles" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157368 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d2b0ef8-641f-462a-be15-8ecb8f06db6a" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157377 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f460c3a-18c9-46d2-b590-f86729f598a5" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.157386 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="350d3b8d-c680-4373-87a8-e1c22d302c64" containerName="registry-server" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.159842 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.175749 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fcz4s"] Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.269967 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7jsg\" (UniqueName: \"kubernetes.io/projected/749f8c14-287b-497c-ad5e-cdc9c45b30fe-kube-api-access-f7jsg\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.270023 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.270064 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.372507 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7jsg\" (UniqueName: \"kubernetes.io/projected/749f8c14-287b-497c-ad5e-cdc9c45b30fe-kube-api-access-f7jsg\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.373671 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.373710 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.375585 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.397760 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7jsg\" (UniqueName: \"kubernetes.io/projected/749f8c14-287b-497c-ad5e-cdc9c45b30fe-kube-api-access-f7jsg\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.397834 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/749f8c14-287b-497c-ad5e-cdc9c45b30fe-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-fcz4s\" (UID: \"749f8c14-287b-497c-ad5e-cdc9c45b30fe\") " pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.477474 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.608930 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.655410 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.696000 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779463 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content\") pod \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779511 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5scr\" (UniqueName: \"kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr\") pod \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779557 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities\") pod \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\" (UID: \"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779577 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content\") pod \"a174e67d-8201-4967-88d2-2c47b6976df3\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779595 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgxph\" (UniqueName: \"kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph\") pod \"a174e67d-8201-4967-88d2-2c47b6976df3\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.779670 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities\") pod \"a174e67d-8201-4967-88d2-2c47b6976df3\" (UID: \"a174e67d-8201-4967-88d2-2c47b6976df3\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.782920 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities" (OuterVolumeSpecName: "utilities") pod "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" (UID: "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.783799 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities" (OuterVolumeSpecName: "utilities") pod "a174e67d-8201-4967-88d2-2c47b6976df3" (UID: "a174e67d-8201-4967-88d2-2c47b6976df3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.806432 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph" (OuterVolumeSpecName: "kube-api-access-lgxph") pod "a174e67d-8201-4967-88d2-2c47b6976df3" (UID: "a174e67d-8201-4967-88d2-2c47b6976df3"). InnerVolumeSpecName "kube-api-access-lgxph". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.808899 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr" (OuterVolumeSpecName: "kube-api-access-w5scr") pod "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" (UID: "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f"). InnerVolumeSpecName "kube-api-access-w5scr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.870911 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" (UID: "91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.881427 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bq25j\" (UniqueName: \"kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j\") pod \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.881504 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities\") pod \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.881593 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content\") pod \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\" (UID: \"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd\") " Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.882050 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.882065 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5scr\" (UniqueName: \"kubernetes.io/projected/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-kube-api-access-w5scr\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.882078 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.882089 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgxph\" (UniqueName: \"kubernetes.io/projected/a174e67d-8201-4967-88d2-2c47b6976df3-kube-api-access-lgxph\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.882099 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.883191 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities" (OuterVolumeSpecName: "utilities") pod "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" (UID: "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.888812 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j" (OuterVolumeSpecName: "kube-api-access-bq25j") pod "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" (UID: "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd"). InnerVolumeSpecName "kube-api-access-bq25j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.891700 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a174e67d-8201-4967-88d2-2c47b6976df3" (UID: "a174e67d-8201-4967-88d2-2c47b6976df3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.985448 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bq25j\" (UniqueName: \"kubernetes.io/projected/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-kube-api-access-bq25j\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.985505 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:12 crc kubenswrapper[4831]: I1124 08:19:12.985516 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a174e67d-8201-4967-88d2-2c47b6976df3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.077991 4831 generic.go:334] "Generic (PLEG): container finished" podID="a174e67d-8201-4967-88d2-2c47b6976df3" containerID="3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb" exitCode=0 Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.078156 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerDied","Data":"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.078198 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfvtn" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.078230 4831 scope.go:117] "RemoveContainer" containerID="3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.078207 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfvtn" event={"ID":"a174e67d-8201-4967-88d2-2c47b6976df3","Type":"ContainerDied","Data":"4847fe85585e20d38263014777e5581c9a4347b6e8e51f9a5ff4e4e596048e45"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.091521 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.105970 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" (UID: "7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.124128 4831 generic.go:334] "Generic (PLEG): container finished" podID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerID="683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6" exitCode=0 Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.124241 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerDied","Data":"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.124273 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kd4td" event={"ID":"7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd","Type":"ContainerDied","Data":"2c88a55fe586d5bdc75add55371be7b9896a00f2eb4c3e2b3130d6642ef1653d"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.124451 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kd4td" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.142758 4831 generic.go:334] "Generic (PLEG): container finished" podID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerID="59145cdabd2e4dfea08b2c82ff15534d7a0af711498acfbaeabc46106bfd1eeb" exitCode=0 Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.142884 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" event={"ID":"ff236b0b-a353-4f4e-9d87-06739909b3c2","Type":"ContainerDied","Data":"59145cdabd2e4dfea08b2c82ff15534d7a0af711498acfbaeabc46106bfd1eeb"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.149557 4831 scope.go:117] "RemoveContainer" containerID="702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.150398 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.157295 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfvtn"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.171137 4831 generic.go:334] "Generic (PLEG): container finished" podID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerID="60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1" exitCode=0 Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.171231 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerDied","Data":"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.171260 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tks88" event={"ID":"91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f","Type":"ContainerDied","Data":"76bc74ab6490b03f1e678a0018d2bcdd1b1f8b963985d3bb68e92ff2b89dc6a0"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.171378 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tks88" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.188663 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.192144 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.193216 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kd4td"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.214065 4831 scope.go:117] "RemoveContainer" containerID="5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.244574 4831 generic.go:334] "Generic (PLEG): container finished" podID="184380c6-d131-4447-9549-f93708be879c" containerID="7080ffc57f89eed628c3030ab52a09d2966b6c64e98dc17a84add7d601ec1f50" exitCode=0 Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.244642 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerDied","Data":"7080ffc57f89eed628c3030ab52a09d2966b6c64e98dc17a84add7d601ec1f50"} Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.259603 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.273647 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tks88"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.285987 4831 scope.go:117] "RemoveContainer" containerID="3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.290236 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb\": container with ID starting with 3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb not found: ID does not exist" containerID="3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.290300 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb"} err="failed to get container status \"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb\": rpc error: code = NotFound desc = could not find container \"3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb\": container with ID starting with 3348db8299f2015affe247cd1d402a47b3bb32764b63fde429617055b4d2e7fb not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.290355 4831 scope.go:117] "RemoveContainer" containerID="702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.293586 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce\": container with ID starting with 702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce not found: ID does not exist" containerID="702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.293613 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce"} err="failed to get container status \"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce\": rpc error: code = NotFound desc = could not find container \"702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce\": container with ID starting with 702f76122d53fa9763b04d36c4f1cbb92383972b21386b5ff3dc05ea5a0a94ce not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.293628 4831 scope.go:117] "RemoveContainer" containerID="5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.297807 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c\": container with ID starting with 5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c not found: ID does not exist" containerID="5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.298089 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c"} err="failed to get container status \"5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c\": rpc error: code = NotFound desc = could not find container \"5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c\": container with ID starting with 5e5f3eee637d0829a2d5abf3f09331bde868416ff8017dd264fb5e499907ab6c not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.298111 4831 scope.go:117] "RemoveContainer" containerID="683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.313845 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-fcz4s"] Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.357595 4831 scope.go:117] "RemoveContainer" containerID="11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.405769 4831 scope.go:117] "RemoveContainer" containerID="1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.456736 4831 scope.go:117] "RemoveContainer" containerID="683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.460468 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6\": container with ID starting with 683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6 not found: ID does not exist" containerID="683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.460505 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6"} err="failed to get container status \"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6\": rpc error: code = NotFound desc = could not find container \"683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6\": container with ID starting with 683ffabe9b5883904e357f7f4cf265f84f8c77a15178a2315551377d8821eaf6 not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.460531 4831 scope.go:117] "RemoveContainer" containerID="11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.461654 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95\": container with ID starting with 11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95 not found: ID does not exist" containerID="11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.461684 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95"} err="failed to get container status \"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95\": rpc error: code = NotFound desc = could not find container \"11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95\": container with ID starting with 11dad2720d19a84786c07fc7fb846ad4fc7757bf67a9a350bdf5dbbe99754d95 not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.461699 4831 scope.go:117] "RemoveContainer" containerID="1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.466433 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c\": container with ID starting with 1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c not found: ID does not exist" containerID="1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.466462 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c"} err="failed to get container status \"1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c\": rpc error: code = NotFound desc = could not find container \"1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c\": container with ID starting with 1fcca5f79aadb65afcd0c67ea61058490d2e6858002b9315d2d5879c6dc3118c not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.466481 4831 scope.go:117] "RemoveContainer" containerID="60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.495549 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.515855 4831 scope.go:117] "RemoveContainer" containerID="eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.563813 4831 scope.go:117] "RemoveContainer" containerID="ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.609294 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlqv5\" (UniqueName: \"kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5\") pod \"ff236b0b-a353-4f4e-9d87-06739909b3c2\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.609371 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics\") pod \"ff236b0b-a353-4f4e-9d87-06739909b3c2\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.609418 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca\") pod \"ff236b0b-a353-4f4e-9d87-06739909b3c2\" (UID: \"ff236b0b-a353-4f4e-9d87-06739909b3c2\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.610289 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ff236b0b-a353-4f4e-9d87-06739909b3c2" (UID: "ff236b0b-a353-4f4e-9d87-06739909b3c2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.616644 4831 scope.go:117] "RemoveContainer" containerID="60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.621587 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ff236b0b-a353-4f4e-9d87-06739909b3c2" (UID: "ff236b0b-a353-4f4e-9d87-06739909b3c2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.621634 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5" (OuterVolumeSpecName: "kube-api-access-hlqv5") pod "ff236b0b-a353-4f4e-9d87-06739909b3c2" (UID: "ff236b0b-a353-4f4e-9d87-06739909b3c2"). InnerVolumeSpecName "kube-api-access-hlqv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.625030 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1\": container with ID starting with 60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1 not found: ID does not exist" containerID="60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.625078 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1"} err="failed to get container status \"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1\": rpc error: code = NotFound desc = could not find container \"60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1\": container with ID starting with 60e1d6b9e44d3d3f1a6c06b5c18380877c55dd201ed65fd66e9d33b2a8f2f0c1 not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.625116 4831 scope.go:117] "RemoveContainer" containerID="eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.625474 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33\": container with ID starting with eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33 not found: ID does not exist" containerID="eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.625494 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33"} err="failed to get container status \"eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33\": rpc error: code = NotFound desc = could not find container \"eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33\": container with ID starting with eaac545868f2da2639c2a9450b828b5cd4f817b22a5938ced8e90f24eeb71e33 not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.625510 4831 scope.go:117] "RemoveContainer" containerID="ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790" Nov 24 08:19:13 crc kubenswrapper[4831]: E1124 08:19:13.625709 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790\": container with ID starting with ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790 not found: ID does not exist" containerID="ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.625732 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790"} err="failed to get container status \"ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790\": rpc error: code = NotFound desc = could not find container \"ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790\": container with ID starting with ddcb16745c8f0de76f4cc20da43694a51ede3425c4e3ae380f68337a0fe6c790 not found: ID does not exist" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.640334 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710013 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrtm2\" (UniqueName: \"kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2\") pod \"184380c6-d131-4447-9549-f93708be879c\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710156 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities\") pod \"184380c6-d131-4447-9549-f93708be879c\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710214 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content\") pod \"184380c6-d131-4447-9549-f93708be879c\" (UID: \"184380c6-d131-4447-9549-f93708be879c\") " Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710453 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlqv5\" (UniqueName: \"kubernetes.io/projected/ff236b0b-a353-4f4e-9d87-06739909b3c2-kube-api-access-hlqv5\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710477 4831 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.710490 4831 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff236b0b-a353-4f4e-9d87-06739909b3c2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.711426 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities" (OuterVolumeSpecName: "utilities") pod "184380c6-d131-4447-9549-f93708be879c" (UID: "184380c6-d131-4447-9549-f93708be879c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.715273 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2" (OuterVolumeSpecName: "kube-api-access-nrtm2") pod "184380c6-d131-4447-9549-f93708be879c" (UID: "184380c6-d131-4447-9549-f93708be879c"). InnerVolumeSpecName "kube-api-access-nrtm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.766043 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "184380c6-d131-4447-9549-f93708be879c" (UID: "184380c6-d131-4447-9549-f93708be879c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.811932 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.811983 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184380c6-d131-4447-9549-f93708be879c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:13 crc kubenswrapper[4831]: I1124 08:19:13.811996 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrtm2\" (UniqueName: \"kubernetes.io/projected/184380c6-d131-4447-9549-f93708be879c-kube-api-access-nrtm2\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.181482 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230589 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jlh4"] Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230845 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230868 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230882 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230891 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230900 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230907 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230926 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230934 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230944 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230951 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230961 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230970 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.230981 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.230989 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231001 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231008 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231018 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231025 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231036 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231043 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231050 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231056 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231066 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231072 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231080 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231085 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="extract-content" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231092 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231097 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231105 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231110 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.231119 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231124 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="extract-utilities" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231227 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="184380c6-d131-4447-9549-f93708be879c" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231239 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231249 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" containerName="marketplace-operator" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231258 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231267 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.231280 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" containerName="registry-server" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.232009 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.252354 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" event={"ID":"749f8c14-287b-497c-ad5e-cdc9c45b30fe","Type":"ContainerStarted","Data":"41b5455048a6baba0704ef07c69317a80e01046dd66f2b236cee3f113228fdf3"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.252412 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" event={"ID":"749f8c14-287b-497c-ad5e-cdc9c45b30fe","Type":"ContainerStarted","Data":"a914e989f18409e47d87005bcb8094b825f5642b2932e0ba33e8725d1b31f6e4"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.253749 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.255738 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.259072 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.261201 4831 generic.go:334] "Generic (PLEG): container finished" podID="9c997027-92eb-4878-beb2-6fbb939737ce" containerID="347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a" exitCode=0 Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.261268 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerDied","Data":"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.261306 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lkcs" event={"ID":"9c997027-92eb-4878-beb2-6fbb939737ce","Type":"ContainerDied","Data":"64911dce70f50faae47cd7c75af6b9f01727d4b466973ba3dbb9017d2b19d638"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.261343 4831 scope.go:117] "RemoveContainer" containerID="347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.261439 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lkcs" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.268582 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" event={"ID":"ff236b0b-a353-4f4e-9d87-06739909b3c2","Type":"ContainerDied","Data":"32ffe203259cb325cfb641648d94d08649440a35bf94b2ed1de78df30f93bc13"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.268713 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5t9z8" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.274197 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jlh4"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.274433 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fcscl" event={"ID":"184380c6-d131-4447-9549-f93708be879c","Type":"ContainerDied","Data":"eb5ef42dbdc73aa0d6e037edce02184d9b76a582b34ea07be386a9bc5d2adc00"} Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.274545 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fcscl" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.285482 4831 scope.go:117] "RemoveContainer" containerID="51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331236 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities\") pod \"9c997027-92eb-4878-beb2-6fbb939737ce\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331282 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content\") pod \"9c997027-92eb-4878-beb2-6fbb939737ce\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331295 4831 scope.go:117] "RemoveContainer" containerID="8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331427 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qkn4\" (UniqueName: \"kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4\") pod \"9c997027-92eb-4878-beb2-6fbb939737ce\" (UID: \"9c997027-92eb-4878-beb2-6fbb939737ce\") " Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331574 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j88l\" (UniqueName: \"kubernetes.io/projected/dc07c012-a5d2-4527-a476-87ff3aae92d8-kube-api-access-5j88l\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331670 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-utilities\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.331708 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-catalog-content\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.334971 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities" (OuterVolumeSpecName: "utilities") pod "9c997027-92eb-4878-beb2-6fbb939737ce" (UID: "9c997027-92eb-4878-beb2-6fbb939737ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.340451 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4" (OuterVolumeSpecName: "kube-api-access-9qkn4") pod "9c997027-92eb-4878-beb2-6fbb939737ce" (UID: "9c997027-92eb-4878-beb2-6fbb939737ce"). InnerVolumeSpecName "kube-api-access-9qkn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.352397 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.354944 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5t9z8"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.358869 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-fcz4s" podStartSLOduration=2.358846114 podStartE2EDuration="2.358846114s" podCreationTimestamp="2025-11-24 08:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:19:14.356765955 +0000 UTC m=+228.231911118" watchObservedRunningTime="2025-11-24 08:19:14.358846114 +0000 UTC m=+228.233991267" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.376242 4831 scope.go:117] "RemoveContainer" containerID="347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.377039 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a\": container with ID starting with 347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a not found: ID does not exist" containerID="347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.377085 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a"} err="failed to get container status \"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a\": rpc error: code = NotFound desc = could not find container \"347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a\": container with ID starting with 347d208e67bbd9538e18d2a2bac3d6715069ac35a9a9334256aa0c96eff9237a not found: ID does not exist" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.377118 4831 scope.go:117] "RemoveContainer" containerID="51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.377615 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2\": container with ID starting with 51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2 not found: ID does not exist" containerID="51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.377764 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2"} err="failed to get container status \"51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2\": rpc error: code = NotFound desc = could not find container \"51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2\": container with ID starting with 51ef6012d2a371f700e9cd893dbc0d9187149e406b8e0b1ca7d62104bf6392d2 not found: ID does not exist" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.377880 4831 scope.go:117] "RemoveContainer" containerID="8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c" Nov 24 08:19:14 crc kubenswrapper[4831]: E1124 08:19:14.378352 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c\": container with ID starting with 8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c not found: ID does not exist" containerID="8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.378378 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c"} err="failed to get container status \"8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c\": rpc error: code = NotFound desc = could not find container \"8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c\": container with ID starting with 8f21f52edf9771ea07fd0e3d52db0e945c240589e39fd9f7b18645f2b4e2a91c not found: ID does not exist" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.378393 4831 scope.go:117] "RemoveContainer" containerID="59145cdabd2e4dfea08b2c82ff15534d7a0af711498acfbaeabc46106bfd1eeb" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.391048 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.404681 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fcscl"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.422646 4831 scope.go:117] "RemoveContainer" containerID="7080ffc57f89eed628c3030ab52a09d2966b6c64e98dc17a84add7d601ec1f50" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.434794 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-utilities\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.434874 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-catalog-content\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.434924 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j88l\" (UniqueName: \"kubernetes.io/projected/dc07c012-a5d2-4527-a476-87ff3aae92d8-kube-api-access-5j88l\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.435021 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qkn4\" (UniqueName: \"kubernetes.io/projected/9c997027-92eb-4878-beb2-6fbb939737ce-kube-api-access-9qkn4\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.435033 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.435997 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-utilities\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.436222 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc07c012-a5d2-4527-a476-87ff3aae92d8-catalog-content\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.454857 4831 scope.go:117] "RemoveContainer" containerID="154792157e9769e87de2c3400c07b30b925fffbf778e45716ac29f24c1cf143f" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.457245 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j88l\" (UniqueName: \"kubernetes.io/projected/dc07c012-a5d2-4527-a476-87ff3aae92d8-kube-api-access-5j88l\") pod \"certified-operators-4jlh4\" (UID: \"dc07c012-a5d2-4527-a476-87ff3aae92d8\") " pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.491648 4831 scope.go:117] "RemoveContainer" containerID="c8516e15e9036b7e7880e11a38ed87a300a6f64a4f6c410ae6c6a24757024e49" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.498620 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c997027-92eb-4878-beb2-6fbb939737ce" (UID: "9c997027-92eb-4878-beb2-6fbb939737ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.535677 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c997027-92eb-4878-beb2-6fbb939737ce-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.588127 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.603515 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.606186 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4lkcs"] Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.810513 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jlh4"] Nov 24 08:19:14 crc kubenswrapper[4831]: W1124 08:19:14.832310 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc07c012_a5d2_4527_a476_87ff3aae92d8.slice/crio-c4fcb1b45fcff0b8463f00d063abbd0f34ad168e40afedfda79028f1164f568e WatchSource:0}: Error finding container c4fcb1b45fcff0b8463f00d063abbd0f34ad168e40afedfda79028f1164f568e: Status 404 returned error can't find the container with id c4fcb1b45fcff0b8463f00d063abbd0f34ad168e40afedfda79028f1164f568e Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.908483 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="184380c6-d131-4447-9549-f93708be879c" path="/var/lib/kubelet/pods/184380c6-d131-4447-9549-f93708be879c/volumes" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.909098 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd" path="/var/lib/kubelet/pods/7a8c0b4d-8d49-4aa5-ab13-e3a4f898f9bd/volumes" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.910148 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f" path="/var/lib/kubelet/pods/91b8f6f3-d0d2-45ef-ae1e-f73b6f7e6f8f/volumes" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.924426 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c997027-92eb-4878-beb2-6fbb939737ce" path="/var/lib/kubelet/pods/9c997027-92eb-4878-beb2-6fbb939737ce/volumes" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.925675 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a174e67d-8201-4967-88d2-2c47b6976df3" path="/var/lib/kubelet/pods/a174e67d-8201-4967-88d2-2c47b6976df3/volumes" Nov 24 08:19:14 crc kubenswrapper[4831]: I1124 08:19:14.927800 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff236b0b-a353-4f4e-9d87-06739909b3c2" path="/var/lib/kubelet/pods/ff236b0b-a353-4f4e-9d87-06739909b3c2/volumes" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.246379 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wm7hx"] Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.248440 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.251656 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.267916 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wm7hx"] Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.295188 4831 generic.go:334] "Generic (PLEG): container finished" podID="dc07c012-a5d2-4527-a476-87ff3aae92d8" containerID="263aea666acc7f55b89bb8a67751439ff986878b06fb25e3cfd89f6efcba5e40" exitCode=0 Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.295300 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jlh4" event={"ID":"dc07c012-a5d2-4527-a476-87ff3aae92d8","Type":"ContainerDied","Data":"263aea666acc7f55b89bb8a67751439ff986878b06fb25e3cfd89f6efcba5e40"} Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.295359 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jlh4" event={"ID":"dc07c012-a5d2-4527-a476-87ff3aae92d8","Type":"ContainerStarted","Data":"c4fcb1b45fcff0b8463f00d063abbd0f34ad168e40afedfda79028f1164f568e"} Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.351474 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lbsg\" (UniqueName: \"kubernetes.io/projected/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-kube-api-access-6lbsg\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.351582 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-utilities\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.351615 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-catalog-content\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.453264 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lbsg\" (UniqueName: \"kubernetes.io/projected/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-kube-api-access-6lbsg\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.453437 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-utilities\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.453474 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-catalog-content\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.454105 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-utilities\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.454150 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-catalog-content\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.484580 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lbsg\" (UniqueName: \"kubernetes.io/projected/8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f-kube-api-access-6lbsg\") pod \"community-operators-wm7hx\" (UID: \"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f\") " pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.567637 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:15 crc kubenswrapper[4831]: I1124 08:19:15.991219 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wm7hx"] Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.308349 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f" containerID="b13599a44ea65fd9b32ad0c239b272962fdccc0c98cd7ee140acc03dee1dee7a" exitCode=0 Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.310141 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wm7hx" event={"ID":"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f","Type":"ContainerDied","Data":"b13599a44ea65fd9b32ad0c239b272962fdccc0c98cd7ee140acc03dee1dee7a"} Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.310201 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wm7hx" event={"ID":"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f","Type":"ContainerStarted","Data":"5985e95d19e61c8276e234f12b2a92607e0555f103731eb44bc16478e9edb144"} Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.623673 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t6xrr"] Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.625139 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.629198 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.654092 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6xrr"] Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.784780 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwp4n\" (UniqueName: \"kubernetes.io/projected/0e1b1f5e-e348-4846-8089-90c6888f51fa-kube-api-access-cwp4n\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.784903 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-catalog-content\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.784966 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-utilities\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.886952 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwp4n\" (UniqueName: \"kubernetes.io/projected/0e1b1f5e-e348-4846-8089-90c6888f51fa-kube-api-access-cwp4n\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.887124 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-catalog-content\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.887195 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-utilities\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.887916 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-catalog-content\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.888015 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e1b1f5e-e348-4846-8089-90c6888f51fa-utilities\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.913892 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwp4n\" (UniqueName: \"kubernetes.io/projected/0e1b1f5e-e348-4846-8089-90c6888f51fa-kube-api-access-cwp4n\") pod \"redhat-marketplace-t6xrr\" (UID: \"0e1b1f5e-e348-4846-8089-90c6888f51fa\") " pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:16 crc kubenswrapper[4831]: I1124 08:19:16.941226 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.317516 4831 generic.go:334] "Generic (PLEG): container finished" podID="dc07c012-a5d2-4527-a476-87ff3aae92d8" containerID="5520a241f4b37024f6e13fe8dfc03216535e48cf671e58ae87e1e6ff84a74b20" exitCode=0 Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.317963 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jlh4" event={"ID":"dc07c012-a5d2-4527-a476-87ff3aae92d8","Type":"ContainerDied","Data":"5520a241f4b37024f6e13fe8dfc03216535e48cf671e58ae87e1e6ff84a74b20"} Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.408702 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6xrr"] Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.630715 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9d6pp"] Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.633577 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.646079 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9d6pp"] Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.646873 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.821642 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsmpf\" (UniqueName: \"kubernetes.io/projected/18b44bbb-ad72-4564-b393-a90ee760fa50-kube-api-access-nsmpf\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.822224 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-utilities\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.822302 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-catalog-content\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.923290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-catalog-content\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.923413 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsmpf\" (UniqueName: \"kubernetes.io/projected/18b44bbb-ad72-4564-b393-a90ee760fa50-kube-api-access-nsmpf\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.923460 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-utilities\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.924265 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-utilities\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.924272 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18b44bbb-ad72-4564-b393-a90ee760fa50-catalog-content\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:17 crc kubenswrapper[4831]: I1124 08:19:17.946672 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsmpf\" (UniqueName: \"kubernetes.io/projected/18b44bbb-ad72-4564-b393-a90ee760fa50-kube-api-access-nsmpf\") pod \"redhat-operators-9d6pp\" (UID: \"18b44bbb-ad72-4564-b393-a90ee760fa50\") " pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.039367 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.339393 4831 generic.go:334] "Generic (PLEG): container finished" podID="0e1b1f5e-e348-4846-8089-90c6888f51fa" containerID="5f065f2d53b9c5d7d42fb346c83c07a02eae0bc999cd0607737ad490308ac002" exitCode=0 Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.339684 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6xrr" event={"ID":"0e1b1f5e-e348-4846-8089-90c6888f51fa","Type":"ContainerDied","Data":"5f065f2d53b9c5d7d42fb346c83c07a02eae0bc999cd0607737ad490308ac002"} Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.339764 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6xrr" event={"ID":"0e1b1f5e-e348-4846-8089-90c6888f51fa","Type":"ContainerStarted","Data":"1fbfa5ba7e798abfb2cea74d3690e5790909fa684b667a4d85b33aee2b951b93"} Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.350038 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f" containerID="9c0d2adab21255fdedf790412aaaab110d6618b77a15ecf072c1c121c45fac8f" exitCode=0 Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.350102 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wm7hx" event={"ID":"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f","Type":"ContainerDied","Data":"9c0d2adab21255fdedf790412aaaab110d6618b77a15ecf072c1c121c45fac8f"} Nov 24 08:19:18 crc kubenswrapper[4831]: I1124 08:19:18.677633 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9d6pp"] Nov 24 08:19:18 crc kubenswrapper[4831]: W1124 08:19:18.690263 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18b44bbb_ad72_4564_b393_a90ee760fa50.slice/crio-485f31a078db1ed0efd4613f47852251efb51fc76ed8f24a09b86d3fd52c2232 WatchSource:0}: Error finding container 485f31a078db1ed0efd4613f47852251efb51fc76ed8f24a09b86d3fd52c2232: Status 404 returned error can't find the container with id 485f31a078db1ed0efd4613f47852251efb51fc76ed8f24a09b86d3fd52c2232 Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.363528 4831 generic.go:334] "Generic (PLEG): container finished" podID="18b44bbb-ad72-4564-b393-a90ee760fa50" containerID="94f21eee69cbd1e0441140d80150a72e36d305d165bf5cf799f5775570ded042" exitCode=0 Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.364197 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d6pp" event={"ID":"18b44bbb-ad72-4564-b393-a90ee760fa50","Type":"ContainerDied","Data":"94f21eee69cbd1e0441140d80150a72e36d305d165bf5cf799f5775570ded042"} Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.364341 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d6pp" event={"ID":"18b44bbb-ad72-4564-b393-a90ee760fa50","Type":"ContainerStarted","Data":"485f31a078db1ed0efd4613f47852251efb51fc76ed8f24a09b86d3fd52c2232"} Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.369522 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wm7hx" event={"ID":"8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f","Type":"ContainerStarted","Data":"1886a6418bb8c9330db32c31ec7254871e16967178df4e46afa646dabf9d36ea"} Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.380074 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jlh4" event={"ID":"dc07c012-a5d2-4527-a476-87ff3aae92d8","Type":"ContainerStarted","Data":"5f5e9ff12faee6ad8e661cd43f0f931a69684a8a580b4f6ad35e9f8411d195ab"} Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.417808 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wm7hx" podStartSLOduration=1.934294071 podStartE2EDuration="4.417787138s" podCreationTimestamp="2025-11-24 08:19:15 +0000 UTC" firstStartedPulling="2025-11-24 08:19:16.310896087 +0000 UTC m=+230.186041240" lastFinishedPulling="2025-11-24 08:19:18.794389154 +0000 UTC m=+232.669534307" observedRunningTime="2025-11-24 08:19:19.41574854 +0000 UTC m=+233.290893713" watchObservedRunningTime="2025-11-24 08:19:19.417787138 +0000 UTC m=+233.292932291" Nov 24 08:19:19 crc kubenswrapper[4831]: I1124 08:19:19.439407 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jlh4" podStartSLOduration=2.52129873 podStartE2EDuration="5.439371578s" podCreationTimestamp="2025-11-24 08:19:14 +0000 UTC" firstStartedPulling="2025-11-24 08:19:15.29845227 +0000 UTC m=+229.173597423" lastFinishedPulling="2025-11-24 08:19:18.216525128 +0000 UTC m=+232.091670271" observedRunningTime="2025-11-24 08:19:19.438251496 +0000 UTC m=+233.313396659" watchObservedRunningTime="2025-11-24 08:19:19.439371578 +0000 UTC m=+233.314516731" Nov 24 08:19:20 crc kubenswrapper[4831]: I1124 08:19:20.387733 4831 generic.go:334] "Generic (PLEG): container finished" podID="0e1b1f5e-e348-4846-8089-90c6888f51fa" containerID="a204d7ccf32bd7d2b9ef3390de7c2409ccd6011b1f96e0f5358da3ae34024b51" exitCode=0 Nov 24 08:19:20 crc kubenswrapper[4831]: I1124 08:19:20.387847 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6xrr" event={"ID":"0e1b1f5e-e348-4846-8089-90c6888f51fa","Type":"ContainerDied","Data":"a204d7ccf32bd7d2b9ef3390de7c2409ccd6011b1f96e0f5358da3ae34024b51"} Nov 24 08:19:21 crc kubenswrapper[4831]: I1124 08:19:21.397809 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6xrr" event={"ID":"0e1b1f5e-e348-4846-8089-90c6888f51fa","Type":"ContainerStarted","Data":"924ff999c7a295aa7b02b87f71ac35257f4777d740239c67f9be7e4be8d54144"} Nov 24 08:19:21 crc kubenswrapper[4831]: I1124 08:19:21.400711 4831 generic.go:334] "Generic (PLEG): container finished" podID="18b44bbb-ad72-4564-b393-a90ee760fa50" containerID="6a5d57018877db5ebea731d17ba43cb4ec6b6d3a85b426478f435d9c82d61140" exitCode=0 Nov 24 08:19:21 crc kubenswrapper[4831]: I1124 08:19:21.400771 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d6pp" event={"ID":"18b44bbb-ad72-4564-b393-a90ee760fa50","Type":"ContainerDied","Data":"6a5d57018877db5ebea731d17ba43cb4ec6b6d3a85b426478f435d9c82d61140"} Nov 24 08:19:21 crc kubenswrapper[4831]: I1124 08:19:21.425193 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t6xrr" podStartSLOduration=2.760342706 podStartE2EDuration="5.42516678s" podCreationTimestamp="2025-11-24 08:19:16 +0000 UTC" firstStartedPulling="2025-11-24 08:19:18.342200867 +0000 UTC m=+232.217346020" lastFinishedPulling="2025-11-24 08:19:21.007024931 +0000 UTC m=+234.882170094" observedRunningTime="2025-11-24 08:19:21.420467445 +0000 UTC m=+235.295612608" watchObservedRunningTime="2025-11-24 08:19:21.42516678 +0000 UTC m=+235.300311933" Nov 24 08:19:24 crc kubenswrapper[4831]: I1124 08:19:24.420145 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9d6pp" event={"ID":"18b44bbb-ad72-4564-b393-a90ee760fa50","Type":"ContainerStarted","Data":"ef52567be7a29269783110ed8a2686c9777e73dd2424ab895853a2fdb3c4e856"} Nov 24 08:19:24 crc kubenswrapper[4831]: I1124 08:19:24.442672 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9d6pp" podStartSLOduration=4.687601247 podStartE2EDuration="7.442651992s" podCreationTimestamp="2025-11-24 08:19:17 +0000 UTC" firstStartedPulling="2025-11-24 08:19:19.365681882 +0000 UTC m=+233.240827035" lastFinishedPulling="2025-11-24 08:19:22.120732627 +0000 UTC m=+235.995877780" observedRunningTime="2025-11-24 08:19:24.440070198 +0000 UTC m=+238.315215381" watchObservedRunningTime="2025-11-24 08:19:24.442651992 +0000 UTC m=+238.317797145" Nov 24 08:19:24 crc kubenswrapper[4831]: I1124 08:19:24.589267 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:24 crc kubenswrapper[4831]: I1124 08:19:24.589351 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:24 crc kubenswrapper[4831]: I1124 08:19:24.632137 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:25 crc kubenswrapper[4831]: I1124 08:19:25.478032 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jlh4" Nov 24 08:19:25 crc kubenswrapper[4831]: I1124 08:19:25.568763 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:25 crc kubenswrapper[4831]: I1124 08:19:25.568979 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:25 crc kubenswrapper[4831]: I1124 08:19:25.614278 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:26 crc kubenswrapper[4831]: I1124 08:19:26.473932 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wm7hx" Nov 24 08:19:26 crc kubenswrapper[4831]: I1124 08:19:26.942154 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:26 crc kubenswrapper[4831]: I1124 08:19:26.942212 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:26 crc kubenswrapper[4831]: I1124 08:19:26.994736 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:27 crc kubenswrapper[4831]: I1124 08:19:27.488997 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t6xrr" Nov 24 08:19:28 crc kubenswrapper[4831]: I1124 08:19:28.040693 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:28 crc kubenswrapper[4831]: I1124 08:19:28.040753 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:29 crc kubenswrapper[4831]: I1124 08:19:29.100361 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9d6pp" podUID="18b44bbb-ad72-4564-b393-a90ee760fa50" containerName="registry-server" probeResult="failure" output=< Nov 24 08:19:29 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:19:29 crc kubenswrapper[4831]: > Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.103969 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.141900 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" containerID="cri-o://4f1974f587e8b3da79701df8c9721ad244c21bc0ed2d9c547747c05038f6ff4b" gracePeriod=15 Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.153342 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9d6pp" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.502448 4831 generic.go:334] "Generic (PLEG): container finished" podID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerID="4f1974f587e8b3da79701df8c9721ad244c21bc0ed2d9c547747c05038f6ff4b" exitCode=0 Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.503135 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" event={"ID":"a5db624a-81de-476b-8df0-24b1c2f0243f","Type":"ContainerDied","Data":"4f1974f587e8b3da79701df8c9721ad244c21bc0ed2d9c547747c05038f6ff4b"} Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.503165 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" event={"ID":"a5db624a-81de-476b-8df0-24b1c2f0243f","Type":"ContainerDied","Data":"a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224"} Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.503177 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a833494d6976aceb5e21db6dc906a42bab895e7d5962815b8e22cac661f95224" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.526437 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.559477 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5566875c7d-w6zbs"] Nov 24 08:19:38 crc kubenswrapper[4831]: E1124 08:19:38.559733 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.559746 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.559866 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" containerName="oauth-openshift" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.560367 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.581468 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5566875c7d-w6zbs"] Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.627922 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.627972 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkm25\" (UniqueName: \"kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628008 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628050 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628084 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628101 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628624 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.628634 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.629798 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.629823 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.629893 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630107 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630161 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630191 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630463 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630510 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630549 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data\") pod \"a5db624a-81de-476b-8df0-24b1c2f0243f\" (UID: \"a5db624a-81de-476b-8df0-24b1c2f0243f\") " Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.630958 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.631422 4831 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.631444 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.631470 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.631527 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.633306 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.638924 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.638952 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.639575 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.639829 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.641085 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.642522 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.643766 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.651906 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25" (OuterVolumeSpecName: "kube-api-access-lkm25") pod "a5db624a-81de-476b-8df0-24b1c2f0243f" (UID: "a5db624a-81de-476b-8df0-24b1c2f0243f"). InnerVolumeSpecName "kube-api-access-lkm25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733017 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733061 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-login\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733088 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733108 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733133 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-session\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733153 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733178 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733201 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqz4p\" (UniqueName: \"kubernetes.io/projected/539f203f-dbde-403a-9dc5-bedd968ca7d1-kube-api-access-sqz4p\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733217 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733240 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733266 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733290 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-policies\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733310 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-error\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733350 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-dir\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733391 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733402 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733411 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733419 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733430 4831 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a5db624a-81de-476b-8df0-24b1c2f0243f-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733438 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733447 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733455 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733463 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkm25\" (UniqueName: \"kubernetes.io/projected/a5db624a-81de-476b-8df0-24b1c2f0243f-kube-api-access-lkm25\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733474 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.733482 4831 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a5db624a-81de-476b-8df0-24b1c2f0243f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.834913 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.834976 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqz4p\" (UniqueName: \"kubernetes.io/projected/539f203f-dbde-403a-9dc5-bedd968ca7d1-kube-api-access-sqz4p\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835010 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835033 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835110 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835149 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-policies\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835181 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-error\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835216 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-dir\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835241 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-login\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835264 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835295 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835347 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835373 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-session\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.835399 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.836107 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-dir\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.836824 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.837085 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.838471 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.839011 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.839275 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-audit-policies\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.839453 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-session\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.839723 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-service-ca\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.840809 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.840811 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-router-certs\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.840835 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-login\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.841253 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-user-template-error\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.844329 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/539f203f-dbde-403a-9dc5-bedd968ca7d1-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.854265 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqz4p\" (UniqueName: \"kubernetes.io/projected/539f203f-dbde-403a-9dc5-bedd968ca7d1-kube-api-access-sqz4p\") pod \"oauth-openshift-5566875c7d-w6zbs\" (UID: \"539f203f-dbde-403a-9dc5-bedd968ca7d1\") " pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:38 crc kubenswrapper[4831]: I1124 08:19:38.873944 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:39 crc kubenswrapper[4831]: I1124 08:19:39.263175 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5566875c7d-w6zbs"] Nov 24 08:19:39 crc kubenswrapper[4831]: W1124 08:19:39.275355 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod539f203f_dbde_403a_9dc5_bedd968ca7d1.slice/crio-204c6e08817b9f248562f7a4b1a9c5c14cddcb91f8ad2b66abd0ac6f96462621 WatchSource:0}: Error finding container 204c6e08817b9f248562f7a4b1a9c5c14cddcb91f8ad2b66abd0ac6f96462621: Status 404 returned error can't find the container with id 204c6e08817b9f248562f7a4b1a9c5c14cddcb91f8ad2b66abd0ac6f96462621 Nov 24 08:19:39 crc kubenswrapper[4831]: I1124 08:19:39.512213 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" event={"ID":"539f203f-dbde-403a-9dc5-bedd968ca7d1","Type":"ContainerStarted","Data":"204c6e08817b9f248562f7a4b1a9c5c14cddcb91f8ad2b66abd0ac6f96462621"} Nov 24 08:19:39 crc kubenswrapper[4831]: I1124 08:19:39.512265 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4nv2q" Nov 24 08:19:39 crc kubenswrapper[4831]: I1124 08:19:39.535476 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:19:39 crc kubenswrapper[4831]: I1124 08:19:39.539561 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4nv2q"] Nov 24 08:19:40 crc kubenswrapper[4831]: I1124 08:19:40.518621 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" event={"ID":"539f203f-dbde-403a-9dc5-bedd968ca7d1","Type":"ContainerStarted","Data":"f414559f84962d7900ca47328fc2c6ffa26796c6df1fd66d4a8d7cac3ad47169"} Nov 24 08:19:40 crc kubenswrapper[4831]: I1124 08:19:40.518900 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:40 crc kubenswrapper[4831]: I1124 08:19:40.526895 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" Nov 24 08:19:40 crc kubenswrapper[4831]: I1124 08:19:40.560718 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5566875c7d-w6zbs" podStartSLOduration=27.560697726 podStartE2EDuration="27.560697726s" podCreationTimestamp="2025-11-24 08:19:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:19:40.540359232 +0000 UTC m=+254.415504395" watchObservedRunningTime="2025-11-24 08:19:40.560697726 +0000 UTC m=+254.435842879" Nov 24 08:19:40 crc kubenswrapper[4831]: I1124 08:19:40.900384 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5db624a-81de-476b-8df0-24b1c2f0243f" path="/var/lib/kubelet/pods/a5db624a-81de-476b-8df0-24b1c2f0243f/volumes" Nov 24 08:21:28 crc kubenswrapper[4831]: I1124 08:21:28.402096 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:21:28 crc kubenswrapper[4831]: I1124 08:21:28.403505 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:21:58 crc kubenswrapper[4831]: I1124 08:21:58.401821 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:21:58 crc kubenswrapper[4831]: I1124 08:21:58.402776 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.401462 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.402244 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.402306 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.403040 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.403103 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac" gracePeriod=600 Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.991916 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac" exitCode=0 Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.992035 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac"} Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.992474 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36"} Nov 24 08:22:28 crc kubenswrapper[4831]: I1124 08:22:28.992511 4831 scope.go:117] "RemoveContainer" containerID="a9b4a4a815593f4ba331f5289c791c4f572d9ed1a04cdcd4d7a04f131f4f4860" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.747210 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-g58jn"] Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.748977 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.770884 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-g58jn"] Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908030 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6m5m\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-kube-api-access-k6m5m\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908390 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-ca-trust-extracted\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908488 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-bound-sa-token\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908614 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-installation-pull-secrets\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908717 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-certificates\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908826 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908903 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-trusted-ca\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.908993 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-tls\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:37 crc kubenswrapper[4831]: I1124 08:22:37.937944 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010779 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-installation-pull-secrets\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010844 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-certificates\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010884 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-trusted-ca\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010906 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-tls\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010938 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6m5m\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-kube-api-access-k6m5m\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010974 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-ca-trust-extracted\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.010994 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-bound-sa-token\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.011790 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-ca-trust-extracted\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.012869 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-trusted-ca\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.013465 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-certificates\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.019416 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-registry-tls\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.027969 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-installation-pull-secrets\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.037216 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-bound-sa-token\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.039289 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6m5m\" (UniqueName: \"kubernetes.io/projected/8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad-kube-api-access-k6m5m\") pod \"image-registry-66df7c8f76-g58jn\" (UID: \"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad\") " pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.066089 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:38 crc kubenswrapper[4831]: I1124 08:22:38.275881 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-g58jn"] Nov 24 08:22:39 crc kubenswrapper[4831]: I1124 08:22:39.061926 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" event={"ID":"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad","Type":"ContainerStarted","Data":"3fd8cd08b092c4fe4d633542b0e8432897a89aef5a9ec0947d0a90eef609339a"} Nov 24 08:22:39 crc kubenswrapper[4831]: I1124 08:22:39.062431 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" event={"ID":"8b6f1b81-ff8f-42c7-8fee-85dcb325c9ad","Type":"ContainerStarted","Data":"54d0351d02a7ed3d9d62b1df05e8b6c0239e2b0eef14ce4f01bef978623a4be3"} Nov 24 08:22:39 crc kubenswrapper[4831]: I1124 08:22:39.062491 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:39 crc kubenswrapper[4831]: I1124 08:22:39.092037 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" podStartSLOduration=2.091978715 podStartE2EDuration="2.091978715s" podCreationTimestamp="2025-11-24 08:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:22:39.082152987 +0000 UTC m=+432.957298160" watchObservedRunningTime="2025-11-24 08:22:39.091978715 +0000 UTC m=+432.967123898" Nov 24 08:22:58 crc kubenswrapper[4831]: I1124 08:22:58.072806 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-g58jn" Nov 24 08:22:58 crc kubenswrapper[4831]: I1124 08:22:58.138514 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.185748 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" podUID="75023939-0385-4a3f-b734-fad2cb5fdc34" containerName="registry" containerID="cri-o://77d8907db75eb830886edf44e19dbb3bfd71262f3c9d716d050398c15c6bc647" gracePeriod=30 Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.368560 4831 generic.go:334] "Generic (PLEG): container finished" podID="75023939-0385-4a3f-b734-fad2cb5fdc34" containerID="77d8907db75eb830886edf44e19dbb3bfd71262f3c9d716d050398c15c6bc647" exitCode=0 Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.368634 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" event={"ID":"75023939-0385-4a3f-b734-fad2cb5fdc34","Type":"ContainerDied","Data":"77d8907db75eb830886edf44e19dbb3bfd71262f3c9d716d050398c15c6bc647"} Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.641552 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717112 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717407 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717467 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717513 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717563 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717607 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8mvr\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717703 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.717836 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets\") pod \"75023939-0385-4a3f-b734-fad2cb5fdc34\" (UID: \"75023939-0385-4a3f-b734-fad2cb5fdc34\") " Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.718817 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.718837 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.719781 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.719831 4831 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.726116 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.727368 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.727480 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.728841 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr" (OuterVolumeSpecName: "kube-api-access-z8mvr") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "kube-api-access-z8mvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.736239 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.758799 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "75023939-0385-4a3f-b734-fad2cb5fdc34" (UID: "75023939-0385-4a3f-b734-fad2cb5fdc34"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.820895 4831 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.821020 4831 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.821074 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8mvr\" (UniqueName: \"kubernetes.io/projected/75023939-0385-4a3f-b734-fad2cb5fdc34-kube-api-access-z8mvr\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.821127 4831 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/75023939-0385-4a3f-b734-fad2cb5fdc34-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:23 crc kubenswrapper[4831]: I1124 08:23:23.821190 4831 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/75023939-0385-4a3f-b734-fad2cb5fdc34-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.379140 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" event={"ID":"75023939-0385-4a3f-b734-fad2cb5fdc34","Type":"ContainerDied","Data":"2f57bb767a01824ae625fabc3b07d5590f059d43341cacae4f9092af86bec74d"} Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.379224 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-6l4dx" Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.379247 4831 scope.go:117] "RemoveContainer" containerID="77d8907db75eb830886edf44e19dbb3bfd71262f3c9d716d050398c15c6bc647" Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.431937 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.439218 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-6l4dx"] Nov 24 08:23:24 crc kubenswrapper[4831]: I1124 08:23:24.907301 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75023939-0385-4a3f-b734-fad2cb5fdc34" path="/var/lib/kubelet/pods/75023939-0385-4a3f-b734-fad2cb5fdc34/volumes" Nov 24 08:24:27 crc kubenswrapper[4831]: I1124 08:24:27.041676 4831 scope.go:117] "RemoveContainer" containerID="4f1974f587e8b3da79701df8c9721ad244c21bc0ed2d9c547747c05038f6ff4b" Nov 24 08:24:28 crc kubenswrapper[4831]: I1124 08:24:28.401584 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:24:28 crc kubenswrapper[4831]: I1124 08:24:28.401962 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.155244 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ppzf6"] Nov 24 08:24:41 crc kubenswrapper[4831]: E1124 08:24:41.156195 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75023939-0385-4a3f-b734-fad2cb5fdc34" containerName="registry" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.156210 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="75023939-0385-4a3f-b734-fad2cb5fdc34" containerName="registry" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.156343 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="75023939-0385-4a3f-b734-fad2cb5fdc34" containerName="registry" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.156837 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.158023 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-clth9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.158875 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-clth9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.165435 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.165568 4831 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-gcjsf" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.166581 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.166657 4831 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7s9fl" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.179051 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ppzf6"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.184789 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-clth9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.193950 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4gh9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.194600 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.197089 4831 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-rts4n" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.214975 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4gh9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.348271 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvc4v\" (UniqueName: \"kubernetes.io/projected/8d75df94-e74a-48b4-b04b-b466f484257d-kube-api-access-vvc4v\") pod \"cert-manager-5b446d88c5-clth9\" (UID: \"8d75df94-e74a-48b4-b04b-b466f484257d\") " pod="cert-manager/cert-manager-5b446d88c5-clth9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.348381 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szdsl\" (UniqueName: \"kubernetes.io/projected/65536f0e-02d6-4cc9-bb72-31cfcc963cdb-kube-api-access-szdsl\") pod \"cert-manager-webhook-5655c58dd6-j4gh9\" (UID: \"65536f0e-02d6-4cc9-bb72-31cfcc963cdb\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.348459 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2848b\" (UniqueName: \"kubernetes.io/projected/c56371b9-cc42-4189-9295-35e64d578865-kube-api-access-2848b\") pod \"cert-manager-cainjector-7f985d654d-ppzf6\" (UID: \"c56371b9-cc42-4189-9295-35e64d578865\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.449984 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvc4v\" (UniqueName: \"kubernetes.io/projected/8d75df94-e74a-48b4-b04b-b466f484257d-kube-api-access-vvc4v\") pod \"cert-manager-5b446d88c5-clth9\" (UID: \"8d75df94-e74a-48b4-b04b-b466f484257d\") " pod="cert-manager/cert-manager-5b446d88c5-clth9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.450507 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szdsl\" (UniqueName: \"kubernetes.io/projected/65536f0e-02d6-4cc9-bb72-31cfcc963cdb-kube-api-access-szdsl\") pod \"cert-manager-webhook-5655c58dd6-j4gh9\" (UID: \"65536f0e-02d6-4cc9-bb72-31cfcc963cdb\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.450802 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2848b\" (UniqueName: \"kubernetes.io/projected/c56371b9-cc42-4189-9295-35e64d578865-kube-api-access-2848b\") pod \"cert-manager-cainjector-7f985d654d-ppzf6\" (UID: \"c56371b9-cc42-4189-9295-35e64d578865\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.473659 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2848b\" (UniqueName: \"kubernetes.io/projected/c56371b9-cc42-4189-9295-35e64d578865-kube-api-access-2848b\") pod \"cert-manager-cainjector-7f985d654d-ppzf6\" (UID: \"c56371b9-cc42-4189-9295-35e64d578865\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.473908 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvc4v\" (UniqueName: \"kubernetes.io/projected/8d75df94-e74a-48b4-b04b-b466f484257d-kube-api-access-vvc4v\") pod \"cert-manager-5b446d88c5-clth9\" (UID: \"8d75df94-e74a-48b4-b04b-b466f484257d\") " pod="cert-manager/cert-manager-5b446d88c5-clth9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.474237 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szdsl\" (UniqueName: \"kubernetes.io/projected/65536f0e-02d6-4cc9-bb72-31cfcc963cdb-kube-api-access-szdsl\") pod \"cert-manager-webhook-5655c58dd6-j4gh9\" (UID: \"65536f0e-02d6-4cc9-bb72-31cfcc963cdb\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.480689 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.490812 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-clth9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.507870 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.780472 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ppzf6"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.798302 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.841759 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-j4gh9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.889809 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-clth9"] Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.977876 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-clth9" event={"ID":"8d75df94-e74a-48b4-b04b-b466f484257d","Type":"ContainerStarted","Data":"a5a52195d5c0fb00c1ad2c1e6db63637aedbda31a600c0445ed54c86789557b6"} Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.978986 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" event={"ID":"c56371b9-cc42-4189-9295-35e64d578865","Type":"ContainerStarted","Data":"00742d59ff9ecc29d7e1054fdc99dcebe409f6460f00443214304d6507d82c08"} Nov 24 08:24:41 crc kubenswrapper[4831]: I1124 08:24:41.980246 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" event={"ID":"65536f0e-02d6-4cc9-bb72-31cfcc963cdb","Type":"ContainerStarted","Data":"1cc8957206eff28d74d1002b2c780c14db7f9c6670113e767463e9d233dd211d"} Nov 24 08:24:45 crc kubenswrapper[4831]: I1124 08:24:45.016340 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-clth9" event={"ID":"8d75df94-e74a-48b4-b04b-b466f484257d","Type":"ContainerStarted","Data":"41f3f856789ef2eeb965ea025675f6351f2200f2bc96f530e8fc2251742b4bca"} Nov 24 08:24:45 crc kubenswrapper[4831]: I1124 08:24:45.018219 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" event={"ID":"c56371b9-cc42-4189-9295-35e64d578865","Type":"ContainerStarted","Data":"f338e126a22594034d8e311804d3794de8727ef2b4d2e21aa5c06c321559f384"} Nov 24 08:24:45 crc kubenswrapper[4831]: I1124 08:24:45.033928 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-clth9" podStartSLOduration=1.288748882 podStartE2EDuration="4.033913699s" podCreationTimestamp="2025-11-24 08:24:41 +0000 UTC" firstStartedPulling="2025-11-24 08:24:41.896806038 +0000 UTC m=+555.771951191" lastFinishedPulling="2025-11-24 08:24:44.641970855 +0000 UTC m=+558.517116008" observedRunningTime="2025-11-24 08:24:45.030950485 +0000 UTC m=+558.906095648" watchObservedRunningTime="2025-11-24 08:24:45.033913699 +0000 UTC m=+558.909058852" Nov 24 08:24:45 crc kubenswrapper[4831]: I1124 08:24:45.051935 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ppzf6" podStartSLOduration=1.207959591 podStartE2EDuration="4.051914182s" podCreationTimestamp="2025-11-24 08:24:41 +0000 UTC" firstStartedPulling="2025-11-24 08:24:41.797962792 +0000 UTC m=+555.673107945" lastFinishedPulling="2025-11-24 08:24:44.641917383 +0000 UTC m=+558.517062536" observedRunningTime="2025-11-24 08:24:45.049012749 +0000 UTC m=+558.924157902" watchObservedRunningTime="2025-11-24 08:24:45.051914182 +0000 UTC m=+558.927059335" Nov 24 08:24:46 crc kubenswrapper[4831]: I1124 08:24:46.027485 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" event={"ID":"65536f0e-02d6-4cc9-bb72-31cfcc963cdb","Type":"ContainerStarted","Data":"96f4b713e1b40ae3f62eda88562038f186cabe6639418159c8f93f9ee1420bb1"} Nov 24 08:24:46 crc kubenswrapper[4831]: I1124 08:24:46.027894 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:46 crc kubenswrapper[4831]: I1124 08:24:46.048198 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" podStartSLOduration=1.059867213 podStartE2EDuration="5.04818175s" podCreationTimestamp="2025-11-24 08:24:41 +0000 UTC" firstStartedPulling="2025-11-24 08:24:41.854072751 +0000 UTC m=+555.729217904" lastFinishedPulling="2025-11-24 08:24:45.842387288 +0000 UTC m=+559.717532441" observedRunningTime="2025-11-24 08:24:46.046808701 +0000 UTC m=+559.921953864" watchObservedRunningTime="2025-11-24 08:24:46.04818175 +0000 UTC m=+559.923326903" Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.514116 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-j4gh9" Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.607916 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hxqjw"] Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608618 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-controller" containerID="cri-o://67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608738 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="nbdb" containerID="cri-o://fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608806 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-node" containerID="cri-o://384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608836 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="northd" containerID="cri-o://481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608876 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-acl-logging" containerID="cri-o://13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608789 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.608990 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="sbdb" containerID="cri-o://320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8" gracePeriod=30 Nov 24 08:24:51 crc kubenswrapper[4831]: I1124 08:24:51.684516 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" containerID="cri-o://3e033d089aeadb785bd84f8d5fdbeffd332301028b5ac8c474e9500b26d4932c" gracePeriod=30 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.072464 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovnkube-controller/3.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.074285 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-acl-logging/0.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.074823 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-controller/0.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075350 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"3e033d089aeadb785bd84f8d5fdbeffd332301028b5ac8c474e9500b26d4932c"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075426 4831 scope.go:117] "RemoveContainer" containerID="1c0c05f29152cd13211da54e6c9e921312677705e13822f3745ff81e0c5ceb8f" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075346 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="3e033d089aeadb785bd84f8d5fdbeffd332301028b5ac8c474e9500b26d4932c" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075488 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075498 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075507 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075518 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075530 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036" exitCode=0 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075537 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8" exitCode=143 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075551 4831 generic.go:334] "Generic (PLEG): container finished" podID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerID="67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed" exitCode=143 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075560 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075604 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075617 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075626 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075636 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075644 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075654 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075664 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" event={"ID":"77826b29-ac89-4b2a-afd9-746c0b1c13ff","Type":"ContainerDied","Data":"b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.075676 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7982aa2d32329a72b9c6167f345228f362988e3adba4134d23f7fb3efae829c" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.078162 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/2.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.078659 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/1.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.078701 4831 generic.go:334] "Generic (PLEG): container finished" podID="89aea50f-21a1-4854-a107-6e38c780166d" containerID="efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac" exitCode=2 Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.078727 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerDied","Data":"efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac"} Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.079187 4831 scope.go:117] "RemoveContainer" containerID="efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.079469 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k6hd5_openshift-multus(89aea50f-21a1-4854-a107-6e38c780166d)\"" pod="openshift-multus/multus-k6hd5" podUID="89aea50f-21a1-4854-a107-6e38c780166d" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.108219 4831 scope.go:117] "RemoveContainer" containerID="a710f3e15a082ec0c2366a5f2d5de5f232b3a81bd7e9fccbcc95aca5f311ae0d" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.118467 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-acl-logging/0.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.119037 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-controller/0.log" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.119783 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.176747 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-b7npl"] Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.176990 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177007 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177016 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177022 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177030 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-node" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177038 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-node" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177051 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="nbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177059 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="nbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177067 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="sbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177074 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="sbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177082 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177090 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177098 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177103 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177110 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="northd" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177116 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="northd" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177125 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177131 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177139 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177144 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177153 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kubecfg-setup" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177160 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kubecfg-setup" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177171 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-acl-logging" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177179 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-acl-logging" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177265 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-node" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177273 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177281 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177294 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="nbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177301 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="kube-rbac-proxy-ovn-metrics" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177309 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177337 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="sbdb" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177346 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="northd" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177353 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovn-acl-logging" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177362 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: E1124 08:24:52.177470 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177479 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177609 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.177621 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" containerName="ovnkube-controller" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.179502 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195236 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-systemd-units\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195283 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-ovn\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195303 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-config\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195406 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195454 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195477 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5kgp\" (UniqueName: \"kubernetes.io/projected/24ebdb49-9e3c-4ed9-9921-819c254fdf74-kube-api-access-j5kgp\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195527 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-etc-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195556 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195589 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-node-log\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195638 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovn-node-metrics-cert\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195673 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-script-lib\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195698 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-systemd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195745 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-log-socket\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195821 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-netns\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195843 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-kubelet\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195859 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-var-lib-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195881 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-bin\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195916 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-netd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.195953 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-slash\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.196005 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-env-overrides\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297002 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297137 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297178 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297221 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297254 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297252 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297303 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297430 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297476 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297510 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297555 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297592 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8gjg\" (UniqueName: \"kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297640 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297346 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297665 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297365 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297607 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297695 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297711 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log" (OuterVolumeSpecName: "node-log") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297651 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297677 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297757 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297781 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297811 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297814 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297837 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297854 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297862 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297886 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297912 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297930 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket" (OuterVolumeSpecName: "log-socket") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297949 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297955 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.297984 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash\") pod \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\" (UID: \"77826b29-ac89-4b2a-afd9-746c0b1c13ff\") " Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298248 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-netns\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298287 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-kubelet\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298345 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-var-lib-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298384 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-bin\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298415 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-netd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298450 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-slash\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298493 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-env-overrides\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298529 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-systemd-units\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298566 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-ovn\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298608 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298643 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-config\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298683 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298699 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-bin\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298723 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5kgp\" (UniqueName: \"kubernetes.io/projected/24ebdb49-9e3c-4ed9-9921-819c254fdf74-kube-api-access-j5kgp\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298753 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298776 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-etc-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298810 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298831 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-kubelet\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298843 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-node-log\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298869 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-var-lib-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298897 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-slash\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298897 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovn-node-metrics-cert\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298903 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-cni-netd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298942 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-script-lib\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298976 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-systemd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299004 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-etc-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298778 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298794 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash" (OuterVolumeSpecName: "host-slash") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299019 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-log-socket\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299071 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-log-socket\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299117 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299235 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-env-overrides\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298815 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-netns\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.298979 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-openvswitch\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299310 4831 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299351 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-node-log\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299368 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-systemd\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299386 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-systemd-units\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299399 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-run-ovn\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299412 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24ebdb49-9e3c-4ed9-9921-819c254fdf74-host-run-ovn-kubernetes\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299425 4831 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-node-log\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299436 4831 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299447 4831 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299456 4831 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299465 4831 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299474 4831 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-log-socket\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299484 4831 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299493 4831 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299502 4831 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299510 4831 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299520 4831 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-slash\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299531 4831 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299540 4831 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299548 4831 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/77826b29-ac89-4b2a-afd9-746c0b1c13ff-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299560 4831 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299580 4831 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.299795 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-config\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.300596 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovnkube-script-lib\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.303989 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg" (OuterVolumeSpecName: "kube-api-access-h8gjg") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "kube-api-access-h8gjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.304712 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24ebdb49-9e3c-4ed9-9921-819c254fdf74-ovn-node-metrics-cert\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.305174 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.312622 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "77826b29-ac89-4b2a-afd9-746c0b1c13ff" (UID: "77826b29-ac89-4b2a-afd9-746c0b1c13ff"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.317121 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5kgp\" (UniqueName: \"kubernetes.io/projected/24ebdb49-9e3c-4ed9-9921-819c254fdf74-kube-api-access-j5kgp\") pod \"ovnkube-node-b7npl\" (UID: \"24ebdb49-9e3c-4ed9-9921-819c254fdf74\") " pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.400468 4831 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/77826b29-ac89-4b2a-afd9-746c0b1c13ff-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.400516 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8gjg\" (UniqueName: \"kubernetes.io/projected/77826b29-ac89-4b2a-afd9-746c0b1c13ff-kube-api-access-h8gjg\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.400525 4831 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/77826b29-ac89-4b2a-afd9-746c0b1c13ff-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 24 08:24:52 crc kubenswrapper[4831]: I1124 08:24:52.492049 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.103139 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-acl-logging/0.log" Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.104228 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hxqjw_77826b29-ac89-4b2a-afd9-746c0b1c13ff/ovn-controller/0.log" Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.105099 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hxqjw" Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.107955 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/2.log" Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.109753 4831 generic.go:334] "Generic (PLEG): container finished" podID="24ebdb49-9e3c-4ed9-9921-819c254fdf74" containerID="759fa2edec0e7e5b220fedaf7d9070a2ad54f4aafb859580e5def1a4ea299822" exitCode=0 Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.109807 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerDied","Data":"759fa2edec0e7e5b220fedaf7d9070a2ad54f4aafb859580e5def1a4ea299822"} Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.109863 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"076da0513b1d94d4e0863184b1e404d9852382bc581ed034d8b1320e82173a01"} Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.172917 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hxqjw"] Nov 24 08:24:53 crc kubenswrapper[4831]: I1124 08:24:53.177056 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hxqjw"] Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120051 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"32b50f0475a497ee960d90787a52842e9c59496ac512fa2344fb8bff751e7b8b"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120499 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"e7cd9a23777a36af1d7f472e96e228e871eb6a610d571a6c833357c5cd48b65c"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120511 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"2a6d090105515fc50296be02509cf6b3f0e2d1de1d9bb703735648dd89721edd"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120521 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"95d42e9b453a25fe063f7bdb6f6999fde4232fb253651441d67d8084ff8bc75a"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120530 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"1a674016850ca38e01f3a32b6505311d27b5eaec88dd622be2d13d95ad752486"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.120539 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"20b9610e2ec28538f6e6f8ea885c2ee2bffbbccb46e212083c30b986f0744fa5"} Nov 24 08:24:54 crc kubenswrapper[4831]: I1124 08:24:54.903144 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77826b29-ac89-4b2a-afd9-746c0b1c13ff" path="/var/lib/kubelet/pods/77826b29-ac89-4b2a-afd9-746c0b1c13ff/volumes" Nov 24 08:24:56 crc kubenswrapper[4831]: I1124 08:24:56.135878 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"c1579bf504d9372a96fef41d7569805a0a983b874db4cc628b4563b4f8e63685"} Nov 24 08:24:58 crc kubenswrapper[4831]: I1124 08:24:58.401423 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:24:58 crc kubenswrapper[4831]: I1124 08:24:58.402137 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:24:59 crc kubenswrapper[4831]: I1124 08:24:59.160217 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" event={"ID":"24ebdb49-9e3c-4ed9-9921-819c254fdf74","Type":"ContainerStarted","Data":"2b4429f62b25faf791e4b50439b0842dc4d4f9c57c5ded1f66b9cbc87636baaa"} Nov 24 08:24:59 crc kubenswrapper[4831]: I1124 08:24:59.160562 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:24:59 crc kubenswrapper[4831]: I1124 08:24:59.192715 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" podStartSLOduration=7.192692261 podStartE2EDuration="7.192692261s" podCreationTimestamp="2025-11-24 08:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:24:59.191418975 +0000 UTC m=+573.066564158" watchObservedRunningTime="2025-11-24 08:24:59.192692261 +0000 UTC m=+573.067837414" Nov 24 08:24:59 crc kubenswrapper[4831]: I1124 08:24:59.198176 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:25:00 crc kubenswrapper[4831]: I1124 08:25:00.166814 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:25:00 crc kubenswrapper[4831]: I1124 08:25:00.168168 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:25:00 crc kubenswrapper[4831]: I1124 08:25:00.200643 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:25:06 crc kubenswrapper[4831]: I1124 08:25:06.899802 4831 scope.go:117] "RemoveContainer" containerID="efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac" Nov 24 08:25:06 crc kubenswrapper[4831]: E1124 08:25:06.903097 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-k6hd5_openshift-multus(89aea50f-21a1-4854-a107-6e38c780166d)\"" pod="openshift-multus/multus-k6hd5" podUID="89aea50f-21a1-4854-a107-6e38c780166d" Nov 24 08:25:18 crc kubenswrapper[4831]: I1124 08:25:18.894145 4831 scope.go:117] "RemoveContainer" containerID="efd2424f9e5da1e4256d6e5a2a97ed253356f9195595575822052c2034f325ac" Nov 24 08:25:19 crc kubenswrapper[4831]: I1124 08:25:19.309207 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-k6hd5_89aea50f-21a1-4854-a107-6e38c780166d/kube-multus/2.log" Nov 24 08:25:19 crc kubenswrapper[4831]: I1124 08:25:19.310014 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-k6hd5" event={"ID":"89aea50f-21a1-4854-a107-6e38c780166d","Type":"ContainerStarted","Data":"208fa70d5029b9a6ee82a81145baebf486fab9ca112ee527c4fe99ea8c9148d3"} Nov 24 08:25:22 crc kubenswrapper[4831]: I1124 08:25:22.523227 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-b7npl" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.103695 4831 scope.go:117] "RemoveContainer" containerID="320efd8bf4c400e0f8de1f57bf692682067f43bb1814f84ee79863c9c03174f8" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.129197 4831 scope.go:117] "RemoveContainer" containerID="13c6b4552fc65ac087f33f34f63f9f09775dab0a88664d2c38024239a890e7a8" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.147460 4831 scope.go:117] "RemoveContainer" containerID="e7be8902e17a34e0dda123d7e93d76f868cfc3fd03e16d66e9f7a826a170a528" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.163876 4831 scope.go:117] "RemoveContainer" containerID="481cfb855c2df5deb46abb8567792694faace5b87c48ef9474e1b4f755dbf077" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.182109 4831 scope.go:117] "RemoveContainer" containerID="fae80adaff30513bf754cc9fde32a83994659eed9422d61297c80e7bc519a9e2" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.201382 4831 scope.go:117] "RemoveContainer" containerID="7eeb66bfae23fd0afc1436d9aebb39e112c52bfc904da16bdddda1d4d2168cbb" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.220936 4831 scope.go:117] "RemoveContainer" containerID="3e033d089aeadb785bd84f8d5fdbeffd332301028b5ac8c474e9500b26d4932c" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.250296 4831 scope.go:117] "RemoveContainer" containerID="384c355501207852d0b2bef0c625b224a80e70da74976b2a8d996eba74430036" Nov 24 08:25:27 crc kubenswrapper[4831]: I1124 08:25:27.266986 4831 scope.go:117] "RemoveContainer" containerID="67a3c1cf03f104a275fd8be4f5d1ca09fc66c8c75a53ac7ad6ab49dcced918ed" Nov 24 08:25:28 crc kubenswrapper[4831]: I1124 08:25:28.401380 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:25:28 crc kubenswrapper[4831]: I1124 08:25:28.401982 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:25:28 crc kubenswrapper[4831]: I1124 08:25:28.402096 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:25:28 crc kubenswrapper[4831]: I1124 08:25:28.403176 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:25:28 crc kubenswrapper[4831]: I1124 08:25:28.403278 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36" gracePeriod=600 Nov 24 08:25:29 crc kubenswrapper[4831]: I1124 08:25:29.390249 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36" exitCode=0 Nov 24 08:25:29 crc kubenswrapper[4831]: I1124 08:25:29.390296 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36"} Nov 24 08:25:29 crc kubenswrapper[4831]: I1124 08:25:29.390768 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090"} Nov 24 08:25:29 crc kubenswrapper[4831]: I1124 08:25:29.390813 4831 scope.go:117] "RemoveContainer" containerID="bf032c705f43b344bfa1171cdb19c36550c8d3f17a785bd8b0a5867cb40d96ac" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.561543 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp"] Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.563458 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.569982 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.581250 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp"] Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.673479 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.673564 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2cpv\" (UniqueName: \"kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.673746 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.775864 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.775926 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2cpv\" (UniqueName: \"kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.775964 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.776491 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.776594 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.798142 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2cpv\" (UniqueName: \"kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:33 crc kubenswrapper[4831]: I1124 08:25:33.894713 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:34 crc kubenswrapper[4831]: I1124 08:25:34.168723 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp"] Nov 24 08:25:34 crc kubenswrapper[4831]: I1124 08:25:34.425831 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerStarted","Data":"337b70dfdfd0e8d3f1bb2b9fab29e7e95d488e31ce13c6585c6c0bf28ce8b46b"} Nov 24 08:25:34 crc kubenswrapper[4831]: I1124 08:25:34.425876 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerStarted","Data":"33c795f39d0526872ddf9a7f6b955efea691072353b6f28e9c6c5304f970d014"} Nov 24 08:25:35 crc kubenswrapper[4831]: I1124 08:25:35.435835 4831 generic.go:334] "Generic (PLEG): container finished" podID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerID="337b70dfdfd0e8d3f1bb2b9fab29e7e95d488e31ce13c6585c6c0bf28ce8b46b" exitCode=0 Nov 24 08:25:35 crc kubenswrapper[4831]: I1124 08:25:35.435898 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerDied","Data":"337b70dfdfd0e8d3f1bb2b9fab29e7e95d488e31ce13c6585c6c0bf28ce8b46b"} Nov 24 08:25:37 crc kubenswrapper[4831]: I1124 08:25:37.450244 4831 generic.go:334] "Generic (PLEG): container finished" podID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerID="3741174a2bb41fc5334a492acf072e48ae54a4da94a8424267c6306d71272086" exitCode=0 Nov 24 08:25:37 crc kubenswrapper[4831]: I1124 08:25:37.450386 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerDied","Data":"3741174a2bb41fc5334a492acf072e48ae54a4da94a8424267c6306d71272086"} Nov 24 08:25:38 crc kubenswrapper[4831]: I1124 08:25:38.464908 4831 generic.go:334] "Generic (PLEG): container finished" podID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerID="c70a7fce78542addb3d958d82dc732916150f9f1edf6cdf3a007c23592471cda" exitCode=0 Nov 24 08:25:38 crc kubenswrapper[4831]: I1124 08:25:38.464976 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerDied","Data":"c70a7fce78542addb3d958d82dc732916150f9f1edf6cdf3a007c23592471cda"} Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.751916 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.847449 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2cpv\" (UniqueName: \"kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv\") pod \"72058868-a5d9-4a81-a6e7-ec9e95140372\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.847683 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle\") pod \"72058868-a5d9-4a81-a6e7-ec9e95140372\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.847723 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util\") pod \"72058868-a5d9-4a81-a6e7-ec9e95140372\" (UID: \"72058868-a5d9-4a81-a6e7-ec9e95140372\") " Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.848764 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle" (OuterVolumeSpecName: "bundle") pod "72058868-a5d9-4a81-a6e7-ec9e95140372" (UID: "72058868-a5d9-4a81-a6e7-ec9e95140372"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.856698 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv" (OuterVolumeSpecName: "kube-api-access-f2cpv") pod "72058868-a5d9-4a81-a6e7-ec9e95140372" (UID: "72058868-a5d9-4a81-a6e7-ec9e95140372"). InnerVolumeSpecName "kube-api-access-f2cpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.949397 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2cpv\" (UniqueName: \"kubernetes.io/projected/72058868-a5d9-4a81-a6e7-ec9e95140372-kube-api-access-f2cpv\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.949459 4831 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:39 crc kubenswrapper[4831]: I1124 08:25:39.998407 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util" (OuterVolumeSpecName: "util") pod "72058868-a5d9-4a81-a6e7-ec9e95140372" (UID: "72058868-a5d9-4a81-a6e7-ec9e95140372"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:25:40 crc kubenswrapper[4831]: I1124 08:25:40.050700 4831 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/72058868-a5d9-4a81-a6e7-ec9e95140372-util\") on node \"crc\" DevicePath \"\"" Nov 24 08:25:40 crc kubenswrapper[4831]: I1124 08:25:40.483263 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" event={"ID":"72058868-a5d9-4a81-a6e7-ec9e95140372","Type":"ContainerDied","Data":"33c795f39d0526872ddf9a7f6b955efea691072353b6f28e9c6c5304f970d014"} Nov 24 08:25:40 crc kubenswrapper[4831]: I1124 08:25:40.483656 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33c795f39d0526872ddf9a7f6b955efea691072353b6f28e9c6c5304f970d014" Nov 24 08:25:40 crc kubenswrapper[4831]: I1124 08:25:40.483365 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.144928 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-h7x4c"] Nov 24 08:25:45 crc kubenswrapper[4831]: E1124 08:25:45.145521 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="extract" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.145537 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="extract" Nov 24 08:25:45 crc kubenswrapper[4831]: E1124 08:25:45.145552 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="util" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.145560 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="util" Nov 24 08:25:45 crc kubenswrapper[4831]: E1124 08:25:45.145572 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="pull" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.145580 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="pull" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.145691 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="72058868-a5d9-4a81-a6e7-ec9e95140372" containerName="extract" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.146135 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.149981 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.155260 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-h7x4c"] Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.156392 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.156394 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-k8ct6" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.216195 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgk8\" (UniqueName: \"kubernetes.io/projected/641a12f7-d840-452b-9d69-4a38b4bb63d7-kube-api-access-qwgk8\") pod \"nmstate-operator-557fdffb88-h7x4c\" (UID: \"641a12f7-d840-452b-9d69-4a38b4bb63d7\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.317183 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwgk8\" (UniqueName: \"kubernetes.io/projected/641a12f7-d840-452b-9d69-4a38b4bb63d7-kube-api-access-qwgk8\") pod \"nmstate-operator-557fdffb88-h7x4c\" (UID: \"641a12f7-d840-452b-9d69-4a38b4bb63d7\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.339726 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwgk8\" (UniqueName: \"kubernetes.io/projected/641a12f7-d840-452b-9d69-4a38b4bb63d7-kube-api-access-qwgk8\") pod \"nmstate-operator-557fdffb88-h7x4c\" (UID: \"641a12f7-d840-452b-9d69-4a38b4bb63d7\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.466877 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" Nov 24 08:25:45 crc kubenswrapper[4831]: I1124 08:25:45.700041 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-h7x4c"] Nov 24 08:25:46 crc kubenswrapper[4831]: I1124 08:25:46.521654 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" event={"ID":"641a12f7-d840-452b-9d69-4a38b4bb63d7","Type":"ContainerStarted","Data":"61ca52b9a43794bf1c5026e6dfd0399ce5b8005f732cac224b219939de4d5f85"} Nov 24 08:25:48 crc kubenswrapper[4831]: I1124 08:25:48.533953 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" event={"ID":"641a12f7-d840-452b-9d69-4a38b4bb63d7","Type":"ContainerStarted","Data":"9649368bc842d217cc7e7b75073eb84812ca70a820fbbd0d58d4430dd1e24241"} Nov 24 08:25:48 crc kubenswrapper[4831]: I1124 08:25:48.566058 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-h7x4c" podStartSLOduration=1.4350768760000001 podStartE2EDuration="3.566021713s" podCreationTimestamp="2025-11-24 08:25:45 +0000 UTC" firstStartedPulling="2025-11-24 08:25:45.721008996 +0000 UTC m=+619.596154149" lastFinishedPulling="2025-11-24 08:25:47.851953833 +0000 UTC m=+621.727098986" observedRunningTime="2025-11-24 08:25:48.561514955 +0000 UTC m=+622.436660138" watchObservedRunningTime="2025-11-24 08:25:48.566021713 +0000 UTC m=+622.441166906" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.858557 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7"] Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.861018 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.862857 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-br2hb" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.877474 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7"] Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.882002 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j"] Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.883200 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.893881 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.909650 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-jv2tr"] Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.910933 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.915546 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j"] Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.922684 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mw6g\" (UniqueName: \"kubernetes.io/projected/11078905-8b2b-4454-890d-11a2f640e692-kube-api-access-9mw6g\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.922753 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-nmstate-lock\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.922823 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.922907 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl8fz\" (UniqueName: \"kubernetes.io/projected/76c1bc31-0ee2-44c8-a2cf-98d66951075c-kube-api-access-zl8fz\") pod \"nmstate-metrics-5dcf9c57c5-qpht7\" (UID: \"76c1bc31-0ee2-44c8-a2cf-98d66951075c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.922937 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9g6t\" (UniqueName: \"kubernetes.io/projected/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-kube-api-access-n9g6t\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.923055 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-ovs-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:53 crc kubenswrapper[4831]: I1124 08:25:53.923083 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-dbus-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024436 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl8fz\" (UniqueName: \"kubernetes.io/projected/76c1bc31-0ee2-44c8-a2cf-98d66951075c-kube-api-access-zl8fz\") pod \"nmstate-metrics-5dcf9c57c5-qpht7\" (UID: \"76c1bc31-0ee2-44c8-a2cf-98d66951075c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024500 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9g6t\" (UniqueName: \"kubernetes.io/projected/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-kube-api-access-n9g6t\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024550 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-ovs-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024599 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-dbus-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024652 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mw6g\" (UniqueName: \"kubernetes.io/projected/11078905-8b2b-4454-890d-11a2f640e692-kube-api-access-9mw6g\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024676 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-nmstate-lock\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.024708 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:54 crc kubenswrapper[4831]: E1124 08:25:54.024850 4831 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Nov 24 08:25:54 crc kubenswrapper[4831]: E1124 08:25:54.024918 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair podName:309f3ff6-5b90-44e6-97d5-4e3fbb794b75 nodeName:}" failed. No retries permitted until 2025-11-24 08:25:54.524895456 +0000 UTC m=+628.400040609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair") pod "nmstate-webhook-6b89b748d8-9wd4j" (UID: "309f3ff6-5b90-44e6-97d5-4e3fbb794b75") : secret "openshift-nmstate-webhook" not found Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.025042 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-nmstate-lock\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.025142 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-dbus-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.025204 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11078905-8b2b-4454-890d-11a2f640e692-ovs-socket\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.032822 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.033716 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.038989 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.039201 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.039376 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-pznsj" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.064251 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9g6t\" (UniqueName: \"kubernetes.io/projected/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-kube-api-access-n9g6t\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.065234 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mw6g\" (UniqueName: \"kubernetes.io/projected/11078905-8b2b-4454-890d-11a2f640e692-kube-api-access-9mw6g\") pod \"nmstate-handler-jv2tr\" (UID: \"11078905-8b2b-4454-890d-11a2f640e692\") " pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.068747 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl8fz\" (UniqueName: \"kubernetes.io/projected/76c1bc31-0ee2-44c8-a2cf-98d66951075c-kube-api-access-zl8fz\") pod \"nmstate-metrics-5dcf9c57c5-qpht7\" (UID: \"76c1bc31-0ee2-44c8-a2cf-98d66951075c\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.072673 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.126008 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz576\" (UniqueName: \"kubernetes.io/projected/5271f267-b380-4526-b535-444b888d87ec-kube-api-access-gz576\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.126082 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5271f267-b380-4526-b535-444b888d87ec-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.126099 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5271f267-b380-4526-b535-444b888d87ec-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.179917 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.226855 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz576\" (UniqueName: \"kubernetes.io/projected/5271f267-b380-4526-b535-444b888d87ec-kube-api-access-gz576\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.226902 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5271f267-b380-4526-b535-444b888d87ec-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.226919 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5271f267-b380-4526-b535-444b888d87ec-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.227821 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5271f267-b380-4526-b535-444b888d87ec-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.232076 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5271f267-b380-4526-b535-444b888d87ec-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.234484 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.266351 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-54c9bfc7b6-s8dk4"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.267539 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.268585 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz576\" (UniqueName: \"kubernetes.io/projected/5271f267-b380-4526-b535-444b888d87ec-kube-api-access-gz576\") pod \"nmstate-console-plugin-5874bd7bc5-cq8tq\" (UID: \"5271f267-b380-4526-b535-444b888d87ec\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.276816 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54c9bfc7b6-s8dk4"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328107 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmsvx\" (UniqueName: \"kubernetes.io/projected/33f82d90-f850-444a-9c08-4e371490ae01-kube-api-access-rmsvx\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328147 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-service-ca\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328174 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328196 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-console-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328232 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-oauth-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328305 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-trusted-ca-bundle\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.328360 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-oauth-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.351948 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429608 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmsvx\" (UniqueName: \"kubernetes.io/projected/33f82d90-f850-444a-9c08-4e371490ae01-kube-api-access-rmsvx\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429644 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-service-ca\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429670 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429692 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-console-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429710 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-oauth-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429743 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-trusted-ca-bundle\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.429764 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-oauth-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.432187 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-oauth-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.432621 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-trusted-ca-bundle\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.434868 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-console-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.438096 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-serving-cert\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.438523 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/33f82d90-f850-444a-9c08-4e371490ae01-service-ca\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.440962 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/33f82d90-f850-444a-9c08-4e371490ae01-console-oauth-config\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.449067 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmsvx\" (UniqueName: \"kubernetes.io/projected/33f82d90-f850-444a-9c08-4e371490ae01-kube-api-access-rmsvx\") pod \"console-54c9bfc7b6-s8dk4\" (UID: \"33f82d90-f850-444a-9c08-4e371490ae01\") " pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.467083 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.530260 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.534538 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/309f3ff6-5b90-44e6-97d5-4e3fbb794b75-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-9wd4j\" (UID: \"309f3ff6-5b90-44e6-97d5-4e3fbb794b75\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.567073 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq"] Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.567627 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jv2tr" event={"ID":"11078905-8b2b-4454-890d-11a2f640e692","Type":"ContainerStarted","Data":"6789e83e1130dbf50c15b1754ef81fab69fa001935f94a1c6281bd8e406369f4"} Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.568867 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" event={"ID":"76c1bc31-0ee2-44c8-a2cf-98d66951075c","Type":"ContainerStarted","Data":"aa930029042258f19604070c6de1a1d638c6975e1c4f4543ae63ce4c84a5c3f1"} Nov 24 08:25:54 crc kubenswrapper[4831]: W1124 08:25:54.573959 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5271f267_b380_4526_b535_444b888d87ec.slice/crio-84a532f38410fdb87ef52a6688c306f4b8e56071ad35308f211a4c7a1830be2d WatchSource:0}: Error finding container 84a532f38410fdb87ef52a6688c306f4b8e56071ad35308f211a4c7a1830be2d: Status 404 returned error can't find the container with id 84a532f38410fdb87ef52a6688c306f4b8e56071ad35308f211a4c7a1830be2d Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.588085 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.796933 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54c9bfc7b6-s8dk4"] Nov 24 08:25:54 crc kubenswrapper[4831]: W1124 08:25:54.805614 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33f82d90_f850_444a_9c08_4e371490ae01.slice/crio-bdfdd3b7be1d6265a75956188b37ac0447972b3e4df6377242f2ee2d819efac8 WatchSource:0}: Error finding container bdfdd3b7be1d6265a75956188b37ac0447972b3e4df6377242f2ee2d819efac8: Status 404 returned error can't find the container with id bdfdd3b7be1d6265a75956188b37ac0447972b3e4df6377242f2ee2d819efac8 Nov 24 08:25:54 crc kubenswrapper[4831]: I1124 08:25:54.806915 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.093836 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j"] Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.575619 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" event={"ID":"5271f267-b380-4526-b535-444b888d87ec","Type":"ContainerStarted","Data":"84a532f38410fdb87ef52a6688c306f4b8e56071ad35308f211a4c7a1830be2d"} Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.579558 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54c9bfc7b6-s8dk4" event={"ID":"33f82d90-f850-444a-9c08-4e371490ae01","Type":"ContainerStarted","Data":"5dce2e6ceacb3fcae3bcd319556810bbce94efa92c7ccca772d510145a0f79ec"} Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.579597 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54c9bfc7b6-s8dk4" event={"ID":"33f82d90-f850-444a-9c08-4e371490ae01","Type":"ContainerStarted","Data":"bdfdd3b7be1d6265a75956188b37ac0447972b3e4df6377242f2ee2d819efac8"} Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.583196 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" event={"ID":"309f3ff6-5b90-44e6-97d5-4e3fbb794b75","Type":"ContainerStarted","Data":"b814297265f4494cb62f2f8d6a2a34103efda90401c4a2c2781debb8a3a4ea55"} Nov 24 08:25:55 crc kubenswrapper[4831]: I1124 08:25:55.598781 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54c9bfc7b6-s8dk4" podStartSLOduration=1.598764871 podStartE2EDuration="1.598764871s" podCreationTimestamp="2025-11-24 08:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:25:55.593224704 +0000 UTC m=+629.468369867" watchObservedRunningTime="2025-11-24 08:25:55.598764871 +0000 UTC m=+629.473910024" Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.601045 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" event={"ID":"76c1bc31-0ee2-44c8-a2cf-98d66951075c","Type":"ContainerStarted","Data":"0e31b5f017a5b291c7da5a5e43536517bb1d663f0b85d747497618da4764cce9"} Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.603552 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" event={"ID":"5271f267-b380-4526-b535-444b888d87ec","Type":"ContainerStarted","Data":"7538207e72f1530703bb3d145ba699a93350c67ee97f5c415b9fd5e2881d8379"} Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.606501 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" event={"ID":"309f3ff6-5b90-44e6-97d5-4e3fbb794b75","Type":"ContainerStarted","Data":"8353f132dfefa8e055be3755814d20b273ba7f04ed2860bb421dff6c08dc972a"} Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.606834 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.624793 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-cq8tq" podStartSLOduration=0.831862702 podStartE2EDuration="3.62476848s" podCreationTimestamp="2025-11-24 08:25:54 +0000 UTC" firstStartedPulling="2025-11-24 08:25:54.577717568 +0000 UTC m=+628.452862721" lastFinishedPulling="2025-11-24 08:25:57.370623346 +0000 UTC m=+631.245768499" observedRunningTime="2025-11-24 08:25:57.620147349 +0000 UTC m=+631.495292512" watchObservedRunningTime="2025-11-24 08:25:57.62476848 +0000 UTC m=+631.499913623" Nov 24 08:25:57 crc kubenswrapper[4831]: I1124 08:25:57.650228 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" podStartSLOduration=2.388668908 podStartE2EDuration="4.650201472s" podCreationTimestamp="2025-11-24 08:25:53 +0000 UTC" firstStartedPulling="2025-11-24 08:25:55.113493437 +0000 UTC m=+628.988638590" lastFinishedPulling="2025-11-24 08:25:57.375026001 +0000 UTC m=+631.250171154" observedRunningTime="2025-11-24 08:25:57.646652061 +0000 UTC m=+631.521797234" watchObservedRunningTime="2025-11-24 08:25:57.650201472 +0000 UTC m=+631.525346625" Nov 24 08:25:58 crc kubenswrapper[4831]: I1124 08:25:58.615873 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-jv2tr" event={"ID":"11078905-8b2b-4454-890d-11a2f640e692","Type":"ContainerStarted","Data":"a2ce58e2f5eae29b8813b05c80033890057e7145bffc19d61ffa0fd7e3529d01"} Nov 24 08:25:58 crc kubenswrapper[4831]: I1124 08:25:58.617104 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:25:58 crc kubenswrapper[4831]: I1124 08:25:58.658154 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-jv2tr" podStartSLOduration=2.605831062 podStartE2EDuration="5.658133153s" podCreationTimestamp="2025-11-24 08:25:53 +0000 UTC" firstStartedPulling="2025-11-24 08:25:54.319295533 +0000 UTC m=+628.194440686" lastFinishedPulling="2025-11-24 08:25:57.371597624 +0000 UTC m=+631.246742777" observedRunningTime="2025-11-24 08:25:58.654038367 +0000 UTC m=+632.529183530" watchObservedRunningTime="2025-11-24 08:25:58.658133153 +0000 UTC m=+632.533278306" Nov 24 08:26:00 crc kubenswrapper[4831]: I1124 08:26:00.629089 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" event={"ID":"76c1bc31-0ee2-44c8-a2cf-98d66951075c","Type":"ContainerStarted","Data":"abcb561f4a866a0b00f98017725490d94948fa8cf8e0e689504c8c31512ef02e"} Nov 24 08:26:00 crc kubenswrapper[4831]: I1124 08:26:00.649732 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-qpht7" podStartSLOduration=2.30389193 podStartE2EDuration="7.649703884s" podCreationTimestamp="2025-11-24 08:25:53 +0000 UTC" firstStartedPulling="2025-11-24 08:25:54.476087643 +0000 UTC m=+628.351232796" lastFinishedPulling="2025-11-24 08:25:59.821899597 +0000 UTC m=+633.697044750" observedRunningTime="2025-11-24 08:26:00.644096425 +0000 UTC m=+634.519241588" watchObservedRunningTime="2025-11-24 08:26:00.649703884 +0000 UTC m=+634.524849077" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.265516 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-jv2tr" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.588660 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.588703 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.594474 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.657939 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54c9bfc7b6-s8dk4" Nov 24 08:26:04 crc kubenswrapper[4831]: I1124 08:26:04.729876 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:26:14 crc kubenswrapper[4831]: I1124 08:26:14.813454 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-9wd4j" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.393039 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m"] Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.395582 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.397552 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.400385 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m"] Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.488748 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.488863 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz7nk\" (UniqueName: \"kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.489019 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.590604 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.590674 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.590707 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz7nk\" (UniqueName: \"kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.591379 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.591748 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.610433 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz7nk\" (UniqueName: \"kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.714094 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.776081 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-9kqsc" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" containerID="cri-o://753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e" gracePeriod=15 Nov 24 08:26:29 crc kubenswrapper[4831]: I1124 08:26:29.951282 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m"] Nov 24 08:26:29 crc kubenswrapper[4831]: W1124 08:26:29.960038 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b86d122_0df8_4bd9_9a18_21413be5e335.slice/crio-025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1 WatchSource:0}: Error finding container 025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1: Status 404 returned error can't find the container with id 025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1 Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.137789 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9kqsc_9d58de3d-6d78-4d9d-92b0-263c45d3ac2b/console/0.log" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.138173 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198530 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198591 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198621 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198687 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198704 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198776 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.198826 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zjfx\" (UniqueName: \"kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx\") pod \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\" (UID: \"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b\") " Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.200107 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.200146 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca" (OuterVolumeSpecName: "service-ca") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.200197 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config" (OuterVolumeSpecName: "console-config") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.200221 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.204846 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.204965 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx" (OuterVolumeSpecName: "kube-api-access-5zjfx") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "kube-api-access-5zjfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.205312 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" (UID: "9d58de3d-6d78-4d9d-92b0-263c45d3ac2b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300713 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zjfx\" (UniqueName: \"kubernetes.io/projected/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-kube-api-access-5zjfx\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300746 4831 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300757 4831 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300766 4831 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300774 4831 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-service-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300784 4831 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.300792 4831 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.858834 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9kqsc_9d58de3d-6d78-4d9d-92b0-263c45d3ac2b/console/0.log" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.858929 4831 generic.go:334] "Generic (PLEG): container finished" podID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerID="753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e" exitCode=2 Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.859034 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kqsc" event={"ID":"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b","Type":"ContainerDied","Data":"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e"} Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.859044 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9kqsc" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.859093 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9kqsc" event={"ID":"9d58de3d-6d78-4d9d-92b0-263c45d3ac2b","Type":"ContainerDied","Data":"cd92280e88715f8cdfd0ca91cab970532d70d929d52cb4011cfde668cd1b2c49"} Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.859118 4831 scope.go:117] "RemoveContainer" containerID="753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.863591 4831 generic.go:334] "Generic (PLEG): container finished" podID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerID="c648d72eb4861f5e3a3b7409c98e4fc0144bbf7cbda81b6e404cc4d711932cf2" exitCode=0 Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.863667 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" event={"ID":"9b86d122-0df8-4bd9-9a18-21413be5e335","Type":"ContainerDied","Data":"c648d72eb4861f5e3a3b7409c98e4fc0144bbf7cbda81b6e404cc4d711932cf2"} Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.863720 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" event={"ID":"9b86d122-0df8-4bd9-9a18-21413be5e335","Type":"ContainerStarted","Data":"025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1"} Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.890383 4831 scope.go:117] "RemoveContainer" containerID="753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e" Nov 24 08:26:30 crc kubenswrapper[4831]: E1124 08:26:30.892031 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e\": container with ID starting with 753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e not found: ID does not exist" containerID="753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.893429 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e"} err="failed to get container status \"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e\": rpc error: code = NotFound desc = could not find container \"753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e\": container with ID starting with 753a6bd0e9eb2600580ebfbf2e9b22be3abda8399f9d8b2a8e6527bd1202225e not found: ID does not exist" Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.923681 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:26:30 crc kubenswrapper[4831]: I1124 08:26:30.928503 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-9kqsc"] Nov 24 08:26:32 crc kubenswrapper[4831]: I1124 08:26:32.906417 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" path="/var/lib/kubelet/pods/9d58de3d-6d78-4d9d-92b0-263c45d3ac2b/volumes" Nov 24 08:26:33 crc kubenswrapper[4831]: I1124 08:26:33.895066 4831 generic.go:334] "Generic (PLEG): container finished" podID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerID="7640239cb35ab0df5a224895ff1170475ce4d1c10f047aa5f473d9dabb79371e" exitCode=0 Nov 24 08:26:33 crc kubenswrapper[4831]: I1124 08:26:33.895131 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" event={"ID":"9b86d122-0df8-4bd9-9a18-21413be5e335","Type":"ContainerDied","Data":"7640239cb35ab0df5a224895ff1170475ce4d1c10f047aa5f473d9dabb79371e"} Nov 24 08:26:34 crc kubenswrapper[4831]: I1124 08:26:34.909232 4831 generic.go:334] "Generic (PLEG): container finished" podID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerID="702f203232d932610bbdc5b24dfd10ac1b93ab102a20494fb5cb486d549f8524" exitCode=0 Nov 24 08:26:34 crc kubenswrapper[4831]: I1124 08:26:34.909356 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" event={"ID":"9b86d122-0df8-4bd9-9a18-21413be5e335","Type":"ContainerDied","Data":"702f203232d932610bbdc5b24dfd10ac1b93ab102a20494fb5cb486d549f8524"} Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.146442 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.284943 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz7nk\" (UniqueName: \"kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk\") pod \"9b86d122-0df8-4bd9-9a18-21413be5e335\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.285075 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util\") pod \"9b86d122-0df8-4bd9-9a18-21413be5e335\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.285174 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle\") pod \"9b86d122-0df8-4bd9-9a18-21413be5e335\" (UID: \"9b86d122-0df8-4bd9-9a18-21413be5e335\") " Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.286478 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle" (OuterVolumeSpecName: "bundle") pod "9b86d122-0df8-4bd9-9a18-21413be5e335" (UID: "9b86d122-0df8-4bd9-9a18-21413be5e335"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.293239 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk" (OuterVolumeSpecName: "kube-api-access-vz7nk") pod "9b86d122-0df8-4bd9-9a18-21413be5e335" (UID: "9b86d122-0df8-4bd9-9a18-21413be5e335"). InnerVolumeSpecName "kube-api-access-vz7nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.302345 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util" (OuterVolumeSpecName: "util") pod "9b86d122-0df8-4bd9-9a18-21413be5e335" (UID: "9b86d122-0df8-4bd9-9a18-21413be5e335"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.386554 4831 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.386889 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz7nk\" (UniqueName: \"kubernetes.io/projected/9b86d122-0df8-4bd9-9a18-21413be5e335-kube-api-access-vz7nk\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.386903 4831 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9b86d122-0df8-4bd9-9a18-21413be5e335-util\") on node \"crc\" DevicePath \"\"" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.933170 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" event={"ID":"9b86d122-0df8-4bd9-9a18-21413be5e335","Type":"ContainerDied","Data":"025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1"} Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.933251 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="025c3630acf91dbf91d3bb913fc67de852c88cc2b51bd340f02626466552bec1" Nov 24 08:26:36 crc kubenswrapper[4831]: I1124 08:26:36.933259 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.310474 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d"] Nov 24 08:26:50 crc kubenswrapper[4831]: E1124 08:26:50.311205 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="util" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311216 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="util" Nov 24 08:26:50 crc kubenswrapper[4831]: E1124 08:26:50.311231 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="extract" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311238 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="extract" Nov 24 08:26:50 crc kubenswrapper[4831]: E1124 08:26:50.311247 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="pull" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311254 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="pull" Nov 24 08:26:50 crc kubenswrapper[4831]: E1124 08:26:50.311266 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311272 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311370 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b86d122-0df8-4bd9-9a18-21413be5e335" containerName="extract" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311379 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d58de3d-6d78-4d9d-92b0-263c45d3ac2b" containerName="console" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.311727 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.314737 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.314832 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.314877 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.315020 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.315227 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7bz2x" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.346952 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d"] Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.453374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-webhook-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.453648 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldrc8\" (UniqueName: \"kubernetes.io/projected/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-kube-api-access-ldrc8\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.453842 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-apiservice-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.555757 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-apiservice-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.555860 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-webhook-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.555900 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldrc8\" (UniqueName: \"kubernetes.io/projected/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-kube-api-access-ldrc8\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.565737 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-apiservice-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.565737 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-webhook-cert\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.602668 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldrc8\" (UniqueName: \"kubernetes.io/projected/7bdd9aa7-6f5e-438e-8646-f05b704b60a6-kube-api-access-ldrc8\") pod \"metallb-operator-controller-manager-7f799f4d7b-m5b2d\" (UID: \"7bdd9aa7-6f5e-438e-8646-f05b704b60a6\") " pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.625885 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.783790 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72"] Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.785073 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.791978 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.792278 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.792460 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5bv89" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.801826 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72"] Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.955616 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d"] Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.960510 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-webhook-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.960552 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-apiservice-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:50 crc kubenswrapper[4831]: I1124 08:26:50.960586 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r92xc\" (UniqueName: \"kubernetes.io/projected/775f3c57-baa8-4e13-8126-eee5194dc2ae-kube-api-access-r92xc\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.008877 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" event={"ID":"7bdd9aa7-6f5e-438e-8646-f05b704b60a6","Type":"ContainerStarted","Data":"43693377aeb35a34e2ac5fd61ad6a750174524960853b0e5773fb37bedd05193"} Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.062119 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-apiservice-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.062155 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-webhook-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.062189 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r92xc\" (UniqueName: \"kubernetes.io/projected/775f3c57-baa8-4e13-8126-eee5194dc2ae-kube-api-access-r92xc\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.068037 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-apiservice-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.068417 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/775f3c57-baa8-4e13-8126-eee5194dc2ae-webhook-cert\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.080077 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r92xc\" (UniqueName: \"kubernetes.io/projected/775f3c57-baa8-4e13-8126-eee5194dc2ae-kube-api-access-r92xc\") pod \"metallb-operator-webhook-server-7d9bd5646d-74m72\" (UID: \"775f3c57-baa8-4e13-8126-eee5194dc2ae\") " pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.125337 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:51 crc kubenswrapper[4831]: I1124 08:26:51.370674 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72"] Nov 24 08:26:52 crc kubenswrapper[4831]: I1124 08:26:52.014543 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" event={"ID":"775f3c57-baa8-4e13-8126-eee5194dc2ae","Type":"ContainerStarted","Data":"357a07b3f03ca7a5f965d53f64144ccb6a0c647b2cbd250c14a8909ceae6ded0"} Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.045086 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" event={"ID":"775f3c57-baa8-4e13-8126-eee5194dc2ae","Type":"ContainerStarted","Data":"1d154390835bf1c6bf94996080db687b6bfe84f2f6e83f5676937c409551bba5"} Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.045705 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.046793 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" event={"ID":"7bdd9aa7-6f5e-438e-8646-f05b704b60a6","Type":"ContainerStarted","Data":"48bfb270ddfeec2a0a976cd66a25da8b976273dd1bc96b29e8b66cc4a82b2bc2"} Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.046969 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.074279 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" podStartSLOduration=1.9324864929999999 podStartE2EDuration="7.074258289s" podCreationTimestamp="2025-11-24 08:26:50 +0000 UTC" firstStartedPulling="2025-11-24 08:26:51.384408252 +0000 UTC m=+685.259553405" lastFinishedPulling="2025-11-24 08:26:56.526180048 +0000 UTC m=+690.401325201" observedRunningTime="2025-11-24 08:26:57.070352239 +0000 UTC m=+690.945497432" watchObservedRunningTime="2025-11-24 08:26:57.074258289 +0000 UTC m=+690.949403452" Nov 24 08:26:57 crc kubenswrapper[4831]: I1124 08:26:57.107550 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" podStartSLOduration=1.600962321 podStartE2EDuration="7.107529048s" podCreationTimestamp="2025-11-24 08:26:50 +0000 UTC" firstStartedPulling="2025-11-24 08:26:50.968925752 +0000 UTC m=+684.844070905" lastFinishedPulling="2025-11-24 08:26:56.475492479 +0000 UTC m=+690.350637632" observedRunningTime="2025-11-24 08:26:57.104836272 +0000 UTC m=+690.979981445" watchObservedRunningTime="2025-11-24 08:26:57.107529048 +0000 UTC m=+690.982674201" Nov 24 08:27:11 crc kubenswrapper[4831]: I1124 08:27:11.133368 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d9bd5646d-74m72" Nov 24 08:27:28 crc kubenswrapper[4831]: I1124 08:27:28.401573 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:27:28 crc kubenswrapper[4831]: I1124 08:27:28.402289 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:27:30 crc kubenswrapper[4831]: I1124 08:27:30.629698 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7f799f4d7b-m5b2d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.451122 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jmst2"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.454054 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.461973 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.462100 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6kfms" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.466003 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.472161 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.472819 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.478494 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.497761 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515529 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqs6l\" (UniqueName: \"kubernetes.io/projected/24e6b9b3-4e9b-4410-bdac-1e359d01375a-kube-api-access-zqs6l\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515590 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzzpl\" (UniqueName: \"kubernetes.io/projected/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-kube-api-access-bzzpl\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515611 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-sockets\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515670 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24e6b9b3-4e9b-4410-bdac-1e359d01375a-cert\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515689 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-startup\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515707 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515722 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-reloader\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515743 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.515758 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-conf\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.559041 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-slgk5"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.560193 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.562619 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.562880 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.563026 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.563210 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-rc6xn" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.580714 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-9d24d"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.581601 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: W1124 08:27:31.584793 4831 reflector.go:561] object-"metallb-system"/"controller-certs-secret": failed to list *v1.Secret: secrets "controller-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 24 08:27:31 crc kubenswrapper[4831]: E1124 08:27:31.584845 4831 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"controller-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"controller-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.601594 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-9d24d"] Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623244 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24e6b9b3-4e9b-4410-bdac-1e359d01375a-cert\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623296 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-startup\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623345 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623365 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-reloader\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623390 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623414 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623432 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-conf\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623460 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623475 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-metrics-certs\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623492 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqs6l\" (UniqueName: \"kubernetes.io/projected/24e6b9b3-4e9b-4410-bdac-1e359d01375a-kube-api-access-zqs6l\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623513 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzzpl\" (UniqueName: \"kubernetes.io/projected/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-kube-api-access-bzzpl\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623530 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95twl\" (UniqueName: \"kubernetes.io/projected/03acb26f-311e-456c-a877-dfb4f9439834-kube-api-access-95twl\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623548 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-sockets\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623571 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcldc\" (UniqueName: \"kubernetes.io/projected/1b0fec8b-1645-468f-acf9-4fdfff8b701b-kube-api-access-hcldc\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623601 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-cert\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623622 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/03acb26f-311e-456c-a877-dfb4f9439834-metallb-excludel2\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: E1124 08:27:31.623730 4831 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Nov 24 08:27:31 crc kubenswrapper[4831]: E1124 08:27:31.623776 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs podName:6aab2a82-5013-4c8e-8f08-3f24223a5e3e nodeName:}" failed. No retries permitted until 2025-11-24 08:27:32.123757312 +0000 UTC m=+725.998902465 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs") pod "frr-k8s-jmst2" (UID: "6aab2a82-5013-4c8e-8f08-3f24223a5e3e") : secret "frr-k8s-certs-secret" not found Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623910 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-reloader\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.623925 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.624218 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-sockets\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.624473 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-startup\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.625292 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-frr-conf\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.630204 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24e6b9b3-4e9b-4410-bdac-1e359d01375a-cert\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.643908 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqs6l\" (UniqueName: \"kubernetes.io/projected/24e6b9b3-4e9b-4410-bdac-1e359d01375a-kube-api-access-zqs6l\") pod \"frr-k8s-webhook-server-6998585d5-q4hq4\" (UID: \"24e6b9b3-4e9b-4410-bdac-1e359d01375a\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.661006 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzzpl\" (UniqueName: \"kubernetes.io/projected/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-kube-api-access-bzzpl\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724698 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-cert\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724747 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/03acb26f-311e-456c-a877-dfb4f9439834-metallb-excludel2\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724810 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724841 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724858 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-metrics-certs\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724880 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95twl\" (UniqueName: \"kubernetes.io/projected/03acb26f-311e-456c-a877-dfb4f9439834-kube-api-access-95twl\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.724905 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcldc\" (UniqueName: \"kubernetes.io/projected/1b0fec8b-1645-468f-acf9-4fdfff8b701b-kube-api-access-hcldc\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: E1124 08:27:31.725025 4831 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 08:27:31 crc kubenswrapper[4831]: E1124 08:27:31.725113 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist podName:03acb26f-311e-456c-a877-dfb4f9439834 nodeName:}" failed. No retries permitted until 2025-11-24 08:27:32.225085191 +0000 UTC m=+726.100230444 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist") pod "speaker-slgk5" (UID: "03acb26f-311e-456c-a877-dfb4f9439834") : secret "metallb-memberlist" not found Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.725633 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/03acb26f-311e-456c-a877-dfb4f9439834-metallb-excludel2\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.727024 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.728216 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-metrics-certs\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.741037 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-cert\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.750138 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95twl\" (UniqueName: \"kubernetes.io/projected/03acb26f-311e-456c-a877-dfb4f9439834-kube-api-access-95twl\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.754454 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcldc\" (UniqueName: \"kubernetes.io/projected/1b0fec8b-1645-468f-acf9-4fdfff8b701b-kube-api-access-hcldc\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:31 crc kubenswrapper[4831]: I1124 08:27:31.785043 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:31.999094 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4"] Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.131965 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.137479 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6aab2a82-5013-4c8e-8f08-3f24223a5e3e-metrics-certs\") pod \"frr-k8s-jmst2\" (UID: \"6aab2a82-5013-4c8e-8f08-3f24223a5e3e\") " pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.233407 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:32 crc kubenswrapper[4831]: E1124 08:27:32.233901 4831 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 08:27:32 crc kubenswrapper[4831]: E1124 08:27:32.234046 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist podName:03acb26f-311e-456c-a877-dfb4f9439834 nodeName:}" failed. No retries permitted until 2025-11-24 08:27:33.234016017 +0000 UTC m=+727.109161330 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist") pod "speaker-slgk5" (UID: "03acb26f-311e-456c-a877-dfb4f9439834") : secret "metallb-memberlist" not found Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.259653 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" event={"ID":"24e6b9b3-4e9b-4410-bdac-1e359d01375a","Type":"ContainerStarted","Data":"8b53b5b1dba754545b14773084c6507e905e74c183e4ea23286e213c184a1f33"} Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.368257 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:32 crc kubenswrapper[4831]: E1124 08:27:32.726050 4831 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: failed to sync secret cache: timed out waiting for the condition Nov 24 08:27:32 crc kubenswrapper[4831]: E1124 08:27:32.726424 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs podName:1b0fec8b-1645-468f-acf9-4fdfff8b701b nodeName:}" failed. No retries permitted until 2025-11-24 08:27:33.226403057 +0000 UTC m=+727.101548210 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs") pod "controller-6c7b4b5f48-9d24d" (UID: "1b0fec8b-1645-468f-acf9-4fdfff8b701b") : failed to sync secret cache: timed out waiting for the condition Nov 24 08:27:32 crc kubenswrapper[4831]: I1124 08:27:32.807079 4831 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.246187 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.246248 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:33 crc kubenswrapper[4831]: E1124 08:27:33.246365 4831 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 24 08:27:33 crc kubenswrapper[4831]: E1124 08:27:33.246452 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist podName:03acb26f-311e-456c-a877-dfb4f9439834 nodeName:}" failed. No retries permitted until 2025-11-24 08:27:35.246430107 +0000 UTC m=+729.121575270 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist") pod "speaker-slgk5" (UID: "03acb26f-311e-456c-a877-dfb4f9439834") : secret "metallb-memberlist" not found Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.252535 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1b0fec8b-1645-468f-acf9-4fdfff8b701b-metrics-certs\") pod \"controller-6c7b4b5f48-9d24d\" (UID: \"1b0fec8b-1645-468f-acf9-4fdfff8b701b\") " pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.265111 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"a253a93ae95bfa5386b60ee634f91951c6f8d726fbeb37c7149ece0528ff7244"} Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.397732 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:33 crc kubenswrapper[4831]: I1124 08:27:33.607104 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-9d24d"] Nov 24 08:27:34 crc kubenswrapper[4831]: I1124 08:27:34.292259 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9d24d" event={"ID":"1b0fec8b-1645-468f-acf9-4fdfff8b701b","Type":"ContainerStarted","Data":"4920c50b77b8817c6f8512c3542daab121743146d32a4a2062e6e2493b3f51d4"} Nov 24 08:27:34 crc kubenswrapper[4831]: I1124 08:27:34.292738 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9d24d" event={"ID":"1b0fec8b-1645-468f-acf9-4fdfff8b701b","Type":"ContainerStarted","Data":"bb7378b1439d07c63497cd88fd57e216efb4fd5d41c6a81e81e0009e5f6ba45f"} Nov 24 08:27:34 crc kubenswrapper[4831]: I1124 08:27:34.292752 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-9d24d" event={"ID":"1b0fec8b-1645-468f-acf9-4fdfff8b701b","Type":"ContainerStarted","Data":"f28ae14cacc261994825169f635592bcdc0b8c509506e1518e581a4a948ef876"} Nov 24 08:27:34 crc kubenswrapper[4831]: I1124 08:27:34.292784 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:34 crc kubenswrapper[4831]: I1124 08:27:34.338210 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-9d24d" podStartSLOduration=3.338191584 podStartE2EDuration="3.338191584s" podCreationTimestamp="2025-11-24 08:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:27:34.332631577 +0000 UTC m=+728.207776740" watchObservedRunningTime="2025-11-24 08:27:34.338191584 +0000 UTC m=+728.213336727" Nov 24 08:27:35 crc kubenswrapper[4831]: I1124 08:27:35.279600 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:35 crc kubenswrapper[4831]: I1124 08:27:35.301078 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/03acb26f-311e-456c-a877-dfb4f9439834-memberlist\") pod \"speaker-slgk5\" (UID: \"03acb26f-311e-456c-a877-dfb4f9439834\") " pod="metallb-system/speaker-slgk5" Nov 24 08:27:35 crc kubenswrapper[4831]: I1124 08:27:35.474697 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-slgk5" Nov 24 08:27:36 crc kubenswrapper[4831]: I1124 08:27:36.319000 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-slgk5" event={"ID":"03acb26f-311e-456c-a877-dfb4f9439834","Type":"ContainerStarted","Data":"e8c3ed85356d2eeeb7f80836fd6208b5ab25acef1af0e91ad7e49b8614f069cc"} Nov 24 08:27:36 crc kubenswrapper[4831]: I1124 08:27:36.319485 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-slgk5" event={"ID":"03acb26f-311e-456c-a877-dfb4f9439834","Type":"ContainerStarted","Data":"d4fb0b5e54cc058165d67e29624b1f8ac66d6253437bd28b8e45b8bff304f95c"} Nov 24 08:27:37 crc kubenswrapper[4831]: I1124 08:27:37.330849 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-slgk5" event={"ID":"03acb26f-311e-456c-a877-dfb4f9439834","Type":"ContainerStarted","Data":"eca1e333766f0c6e8448d139336da6062571ceb7a739a26daa4edb8dcdb96c50"} Nov 24 08:27:37 crc kubenswrapper[4831]: I1124 08:27:37.331239 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-slgk5" Nov 24 08:27:37 crc kubenswrapper[4831]: I1124 08:27:37.354309 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-slgk5" podStartSLOduration=6.354288176 podStartE2EDuration="6.354288176s" podCreationTimestamp="2025-11-24 08:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:27:37.350750396 +0000 UTC m=+731.225895569" watchObservedRunningTime="2025-11-24 08:27:37.354288176 +0000 UTC m=+731.229433329" Nov 24 08:27:41 crc kubenswrapper[4831]: I1124 08:27:41.357741 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" event={"ID":"24e6b9b3-4e9b-4410-bdac-1e359d01375a","Type":"ContainerStarted","Data":"9adec460fc0157a4ede50aae13bdeee3b6624d77ee4f6babf7e9671544d4e2ec"} Nov 24 08:27:41 crc kubenswrapper[4831]: I1124 08:27:41.358030 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:41 crc kubenswrapper[4831]: I1124 08:27:41.359749 4831 generic.go:334] "Generic (PLEG): container finished" podID="6aab2a82-5013-4c8e-8f08-3f24223a5e3e" containerID="5ee1fb34572c3d3b4001cad3c0818c57ffb7bf604305703d3069ad72992b73bd" exitCode=0 Nov 24 08:27:41 crc kubenswrapper[4831]: I1124 08:27:41.359791 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerDied","Data":"5ee1fb34572c3d3b4001cad3c0818c57ffb7bf604305703d3069ad72992b73bd"} Nov 24 08:27:41 crc kubenswrapper[4831]: I1124 08:27:41.398768 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" podStartSLOduration=1.820973615 podStartE2EDuration="10.398738957s" podCreationTimestamp="2025-11-24 08:27:31 +0000 UTC" firstStartedPulling="2025-11-24 08:27:32.011470199 +0000 UTC m=+725.886615352" lastFinishedPulling="2025-11-24 08:27:40.589235541 +0000 UTC m=+734.464380694" observedRunningTime="2025-11-24 08:27:41.377479857 +0000 UTC m=+735.252625050" watchObservedRunningTime="2025-11-24 08:27:41.398738957 +0000 UTC m=+735.273884110" Nov 24 08:27:42 crc kubenswrapper[4831]: I1124 08:27:42.368105 4831 generic.go:334] "Generic (PLEG): container finished" podID="6aab2a82-5013-4c8e-8f08-3f24223a5e3e" containerID="35f63f5179e90634a67002f359055a9ca3bb356f5d951c583581dad3b920e6ac" exitCode=0 Nov 24 08:27:42 crc kubenswrapper[4831]: I1124 08:27:42.368676 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerDied","Data":"35f63f5179e90634a67002f359055a9ca3bb356f5d951c583581dad3b920e6ac"} Nov 24 08:27:43 crc kubenswrapper[4831]: I1124 08:27:43.378196 4831 generic.go:334] "Generic (PLEG): container finished" podID="6aab2a82-5013-4c8e-8f08-3f24223a5e3e" containerID="41f5bb4edbd4f21745caf9ae3a1333cf458b5c7811bd682cf36f6d40831d9083" exitCode=0 Nov 24 08:27:43 crc kubenswrapper[4831]: I1124 08:27:43.378240 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerDied","Data":"41f5bb4edbd4f21745caf9ae3a1333cf458b5c7811bd682cf36f6d40831d9083"} Nov 24 08:27:43 crc kubenswrapper[4831]: I1124 08:27:43.403537 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-9d24d" Nov 24 08:27:44 crc kubenswrapper[4831]: I1124 08:27:44.388585 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"3e18e7599665cc579546dceb4136c1b56891020ee0d7f6879f31a159fece3455"} Nov 24 08:27:44 crc kubenswrapper[4831]: I1124 08:27:44.388642 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"694e2ba51f7bc3807fc78512e0fd0b5c2b094c133b6b3edcd186dc05224e4d4f"} Nov 24 08:27:44 crc kubenswrapper[4831]: I1124 08:27:44.388653 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"b1f86814527a9ec1d7803e0703af0f5438ba6e27b853bd794f2a446ca915478f"} Nov 24 08:27:44 crc kubenswrapper[4831]: I1124 08:27:44.388665 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"2e8762dc8f96318f7c19d4d4894bb30a1723334d6b579774512da447bfb08e52"} Nov 24 08:27:44 crc kubenswrapper[4831]: I1124 08:27:44.388676 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"3d8a515fc26c24fc0efaf1252196478e808d254024ab445d22616327d413cf1a"} Nov 24 08:27:45 crc kubenswrapper[4831]: I1124 08:27:45.402342 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jmst2" event={"ID":"6aab2a82-5013-4c8e-8f08-3f24223a5e3e","Type":"ContainerStarted","Data":"3054ad2409df28a928698b797135ce8bea9a35060087447e828b4d638f67d539"} Nov 24 08:27:45 crc kubenswrapper[4831]: I1124 08:27:45.402774 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:45 crc kubenswrapper[4831]: I1124 08:27:45.433595 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jmst2" podStartSLOduration=6.300786148 podStartE2EDuration="14.433578387s" podCreationTimestamp="2025-11-24 08:27:31 +0000 UTC" firstStartedPulling="2025-11-24 08:27:32.475540331 +0000 UTC m=+726.350685484" lastFinishedPulling="2025-11-24 08:27:40.60833257 +0000 UTC m=+734.483477723" observedRunningTime="2025-11-24 08:27:45.428484163 +0000 UTC m=+739.303629396" watchObservedRunningTime="2025-11-24 08:27:45.433578387 +0000 UTC m=+739.308723540" Nov 24 08:27:47 crc kubenswrapper[4831]: I1124 08:27:47.368867 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:47 crc kubenswrapper[4831]: I1124 08:27:47.410295 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:27:51 crc kubenswrapper[4831]: I1124 08:27:51.792899 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-q4hq4" Nov 24 08:27:55 crc kubenswrapper[4831]: I1124 08:27:55.478972 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-slgk5" Nov 24 08:27:57 crc kubenswrapper[4831]: I1124 08:27:57.488540 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:27:57 crc kubenswrapper[4831]: I1124 08:27:57.489053 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" containerID="cri-o://499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca" gracePeriod=30 Nov 24 08:27:57 crc kubenswrapper[4831]: I1124 08:27:57.597356 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:27:57 crc kubenswrapper[4831]: I1124 08:27:57.597614 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" containerID="cri-o://a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64" gracePeriod=30 Nov 24 08:27:57 crc kubenswrapper[4831]: I1124 08:27:57.918671 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.030334 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert\") pod \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.030392 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles\") pod \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.030417 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config\") pod \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.030443 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22rzm\" (UniqueName: \"kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm\") pod \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.030468 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca\") pod \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\" (UID: \"6a33ab7d-6562-4fbe-8f01-ed62511ff801\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.031144 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca" (OuterVolumeSpecName: "client-ca") pod "6a33ab7d-6562-4fbe-8f01-ed62511ff801" (UID: "6a33ab7d-6562-4fbe-8f01-ed62511ff801"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.031153 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6a33ab7d-6562-4fbe-8f01-ed62511ff801" (UID: "6a33ab7d-6562-4fbe-8f01-ed62511ff801"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.031366 4831 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.031382 4831 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.031974 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config" (OuterVolumeSpecName: "config") pod "6a33ab7d-6562-4fbe-8f01-ed62511ff801" (UID: "6a33ab7d-6562-4fbe-8f01-ed62511ff801"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.040601 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm" (OuterVolumeSpecName: "kube-api-access-22rzm") pod "6a33ab7d-6562-4fbe-8f01-ed62511ff801" (UID: "6a33ab7d-6562-4fbe-8f01-ed62511ff801"). InnerVolumeSpecName "kube-api-access-22rzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.053544 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6a33ab7d-6562-4fbe-8f01-ed62511ff801" (UID: "6a33ab7d-6562-4fbe-8f01-ed62511ff801"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.069934 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.131819 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config\") pod \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.132440 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlhv7\" (UniqueName: \"kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7\") pod \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.132918 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config" (OuterVolumeSpecName: "config") pod "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" (UID: "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133218 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca\") pod \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133228 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" (UID: "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133291 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert\") pod \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\" (UID: \"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b\") " Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133602 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a33ab7d-6562-4fbe-8f01-ed62511ff801-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133624 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22rzm\" (UniqueName: \"kubernetes.io/projected/6a33ab7d-6562-4fbe-8f01-ed62511ff801-kube-api-access-22rzm\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133637 4831 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-client-ca\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133647 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.133658 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a33ab7d-6562-4fbe-8f01-ed62511ff801-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.138136 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7" (OuterVolumeSpecName: "kube-api-access-xlhv7") pod "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" (UID: "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b"). InnerVolumeSpecName "kube-api-access-xlhv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.138134 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" (UID: "7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.235461 4831 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.235515 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlhv7\" (UniqueName: \"kubernetes.io/projected/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b-kube-api-access-xlhv7\") on node \"crc\" DevicePath \"\"" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.340768 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:27:58 crc kubenswrapper[4831]: E1124 08:27:58.340979 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.340991 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: E1124 08:27:58.341009 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.341015 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.341114 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerName="route-controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.341126 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerName="controller-manager" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.341512 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.355798 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.356014 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.358347 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.401258 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.401336 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.438183 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z86gv\" (UniqueName: \"kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv\") pod \"openstack-operator-index-hxwpd\" (UID: \"cf896058-ab59-4f91-bb96-50d059cb9b5b\") " pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.493375 4831 generic.go:334] "Generic (PLEG): container finished" podID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" containerID="499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca" exitCode=0 Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.493415 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.493470 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" event={"ID":"6a33ab7d-6562-4fbe-8f01-ed62511ff801","Type":"ContainerDied","Data":"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca"} Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.493516 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-k2k6z" event={"ID":"6a33ab7d-6562-4fbe-8f01-ed62511ff801","Type":"ContainerDied","Data":"b2cad2a6c46b507020f995c25baaf86bdd3e5867e4c5abc826c9c5535ac679c6"} Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.493535 4831 scope.go:117] "RemoveContainer" containerID="499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.495571 4831 generic.go:334] "Generic (PLEG): container finished" podID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" containerID="a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64" exitCode=0 Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.495591 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" event={"ID":"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b","Type":"ContainerDied","Data":"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64"} Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.495605 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" event={"ID":"7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b","Type":"ContainerDied","Data":"4000e3f846b359b1894a560131b349ebab7fde202cbc5b90aa343fd791c45e1c"} Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.495651 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.514213 4831 scope.go:117] "RemoveContainer" containerID="499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca" Nov 24 08:27:58 crc kubenswrapper[4831]: E1124 08:27:58.516400 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca\": container with ID starting with 499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca not found: ID does not exist" containerID="499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.516436 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca"} err="failed to get container status \"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca\": rpc error: code = NotFound desc = could not find container \"499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca\": container with ID starting with 499cacf3bcfba6b25e083c719ae0cd226580ddc454e594a161eb5d3f20d566ca not found: ID does not exist" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.516456 4831 scope.go:117] "RemoveContainer" containerID="a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.530623 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.540147 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z86gv\" (UniqueName: \"kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv\") pod \"openstack-operator-index-hxwpd\" (UID: \"cf896058-ab59-4f91-bb96-50d059cb9b5b\") " pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.540741 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-k2k6z"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.542746 4831 scope.go:117] "RemoveContainer" containerID="a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64" Nov 24 08:27:58 crc kubenswrapper[4831]: E1124 08:27:58.543835 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64\": container with ID starting with a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64 not found: ID does not exist" containerID="a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.543870 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64"} err="failed to get container status \"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64\": rpc error: code = NotFound desc = could not find container \"a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64\": container with ID starting with a175938142255047f73a2f4b1636a94fc364d18845bd14619652ad895cf65c64 not found: ID does not exist" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.544016 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.546951 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6lg2c"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.557006 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z86gv\" (UniqueName: \"kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv\") pod \"openstack-operator-index-hxwpd\" (UID: \"cf896058-ab59-4f91-bb96-50d059cb9b5b\") " pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.656256 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.862031 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.904765 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a33ab7d-6562-4fbe-8f01-ed62511ff801" path="/var/lib/kubelet/pods/6a33ab7d-6562-4fbe-8f01-ed62511ff801/volumes" Nov 24 08:27:58 crc kubenswrapper[4831]: I1124 08:27:58.909832 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b" path="/var/lib/kubelet/pods/7b4b7eb7-c3ef-4cd2-9d28-6b2f4cbc8c3b/volumes" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.033696 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq"] Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.034533 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.038066 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-797694985b-kp5js"] Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.038235 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.038529 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.038775 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.038947 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.041830 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.042661 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.043201 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047023 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047276 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047441 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047671 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047824 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.047961 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.048516 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq"] Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.053251 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.056102 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-797694985b-kp5js"] Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148192 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r86bm\" (UniqueName: \"kubernetes.io/projected/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-kube-api-access-r86bm\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148253 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-serving-cert\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148279 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-config\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148298 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-client-ca\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148340 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-proxy-ca-bundles\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148366 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-config\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148396 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-client-ca\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148425 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a15aab1-e627-442d-ad9e-2ac477d86a28-serving-cert\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.148456 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9hvg\" (UniqueName: \"kubernetes.io/projected/7a15aab1-e627-442d-ad9e-2ac477d86a28-kube-api-access-n9hvg\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249442 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-serving-cert\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249526 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-config\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249571 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-client-ca\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249629 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-proxy-ca-bundles\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249663 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-config\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249718 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-client-ca\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249744 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a15aab1-e627-442d-ad9e-2ac477d86a28-serving-cert\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249803 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9hvg\" (UniqueName: \"kubernetes.io/projected/7a15aab1-e627-442d-ad9e-2ac477d86a28-kube-api-access-n9hvg\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.249845 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r86bm\" (UniqueName: \"kubernetes.io/projected/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-kube-api-access-r86bm\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.251212 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-client-ca\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.253182 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-config\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.255410 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-config\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.255807 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-client-ca\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.258008 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a15aab1-e627-442d-ad9e-2ac477d86a28-serving-cert\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.263930 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7a15aab1-e627-442d-ad9e-2ac477d86a28-proxy-ca-bundles\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.265382 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9hvg\" (UniqueName: \"kubernetes.io/projected/7a15aab1-e627-442d-ad9e-2ac477d86a28-kube-api-access-n9hvg\") pod \"controller-manager-797694985b-kp5js\" (UID: \"7a15aab1-e627-442d-ad9e-2ac477d86a28\") " pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.270083 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-serving-cert\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.274753 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r86bm\" (UniqueName: \"kubernetes.io/projected/a2ae9cc8-0dcd-4410-a15f-953c7ab418c4-kube-api-access-r86bm\") pod \"route-controller-manager-7c4d5989c4-r9hmq\" (UID: \"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4\") " pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.389329 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.400888 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:27:59 crc kubenswrapper[4831]: I1124 08:27:59.510663 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hxwpd" event={"ID":"cf896058-ab59-4f91-bb96-50d059cb9b5b","Type":"ContainerStarted","Data":"97a66abcba53c74e270c89264f5955bcfaf8a1b58c1f1ad860664e747ec67126"} Nov 24 08:28:00 crc kubenswrapper[4831]: I1124 08:28:00.176570 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq"] Nov 24 08:28:00 crc kubenswrapper[4831]: I1124 08:28:00.264071 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-797694985b-kp5js"] Nov 24 08:28:00 crc kubenswrapper[4831]: W1124 08:28:00.449855 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a15aab1_e627_442d_ad9e_2ac477d86a28.slice/crio-d0a581001c13b9431d9ca37a382f5b3dd6c9fab62afa88643bfa9cbf3793fb73 WatchSource:0}: Error finding container d0a581001c13b9431d9ca37a382f5b3dd6c9fab62afa88643bfa9cbf3793fb73: Status 404 returned error can't find the container with id d0a581001c13b9431d9ca37a382f5b3dd6c9fab62afa88643bfa9cbf3793fb73 Nov 24 08:28:00 crc kubenswrapper[4831]: I1124 08:28:00.521980 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" event={"ID":"7a15aab1-e627-442d-ad9e-2ac477d86a28","Type":"ContainerStarted","Data":"d0a581001c13b9431d9ca37a382f5b3dd6c9fab62afa88643bfa9cbf3793fb73"} Nov 24 08:28:00 crc kubenswrapper[4831]: I1124 08:28:00.523036 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" event={"ID":"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4","Type":"ContainerStarted","Data":"2a377802983cbbadbf8d0f93bdcb53d0c6b62f44e7031107e638d627dcb6fb85"} Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.529975 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" event={"ID":"7a15aab1-e627-442d-ad9e-2ac477d86a28","Type":"ContainerStarted","Data":"08a457b4c3507cd2e64f763ae07ee8477a31ea6e7e92e8c5bec3e81a4ac11216"} Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.530419 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.531926 4831 patch_prober.go:28] interesting pod/controller-manager-797694985b-kp5js container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" start-of-body= Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.531982 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" event={"ID":"a2ae9cc8-0dcd-4410-a15f-953c7ab418c4","Type":"ContainerStarted","Data":"4623851e210716777001434fbf000d9849d947447530e557b397f0f2b88379d5"} Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.532034 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" podUID="7a15aab1-e627-442d-ad9e-2ac477d86a28" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": dial tcp 10.217.0.53:8443: connect: connection refused" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.532186 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.533773 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hxwpd" event={"ID":"cf896058-ab59-4f91-bb96-50d059cb9b5b","Type":"ContainerStarted","Data":"e889962d07fe8db17224e20a58cd5ab6808c16e437ba8fc1aee880f8614e1147"} Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.549246 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" podStartSLOduration=4.549225487 podStartE2EDuration="4.549225487s" podCreationTimestamp="2025-11-24 08:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:28:01.546113049 +0000 UTC m=+755.421258232" watchObservedRunningTime="2025-11-24 08:28:01.549225487 +0000 UTC m=+755.424370640" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.572186 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" podStartSLOduration=4.572165314 podStartE2EDuration="4.572165314s" podCreationTimestamp="2025-11-24 08:27:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:28:01.571731952 +0000 UTC m=+755.446877115" watchObservedRunningTime="2025-11-24 08:28:01.572165314 +0000 UTC m=+755.447310467" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.593661 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hxwpd" podStartSLOduration=1.169444254 podStartE2EDuration="3.59364132s" podCreationTimestamp="2025-11-24 08:27:58 +0000 UTC" firstStartedPulling="2025-11-24 08:27:58.878184408 +0000 UTC m=+752.753329561" lastFinishedPulling="2025-11-24 08:28:01.302381474 +0000 UTC m=+755.177526627" observedRunningTime="2025-11-24 08:28:01.587786534 +0000 UTC m=+755.462931697" watchObservedRunningTime="2025-11-24 08:28:01.59364132 +0000 UTC m=+755.468786473" Nov 24 08:28:01 crc kubenswrapper[4831]: I1124 08:28:01.850187 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7c4d5989c4-r9hmq" Nov 24 08:28:02 crc kubenswrapper[4831]: I1124 08:28:02.371905 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jmst2" Nov 24 08:28:02 crc kubenswrapper[4831]: I1124 08:28:02.542434 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-797694985b-kp5js" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.315199 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.315459 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-hxwpd" podUID="cf896058-ab59-4f91-bb96-50d059cb9b5b" containerName="registry-server" containerID="cri-o://e889962d07fe8db17224e20a58cd5ab6808c16e437ba8fc1aee880f8614e1147" gracePeriod=2 Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.381909 4831 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.556506 4831 generic.go:334] "Generic (PLEG): container finished" podID="cf896058-ab59-4f91-bb96-50d059cb9b5b" containerID="e889962d07fe8db17224e20a58cd5ab6808c16e437ba8fc1aee880f8614e1147" exitCode=0 Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.556545 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hxwpd" event={"ID":"cf896058-ab59-4f91-bb96-50d059cb9b5b","Type":"ContainerDied","Data":"e889962d07fe8db17224e20a58cd5ab6808c16e437ba8fc1aee880f8614e1147"} Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.718773 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xqbxv"] Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.719930 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.721675 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-wcxjl" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.735547 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xqbxv"] Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.838860 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.850072 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w25d8\" (UniqueName: \"kubernetes.io/projected/f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec-kube-api-access-w25d8\") pod \"openstack-operator-index-xqbxv\" (UID: \"f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec\") " pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.951164 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z86gv\" (UniqueName: \"kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv\") pod \"cf896058-ab59-4f91-bb96-50d059cb9b5b\" (UID: \"cf896058-ab59-4f91-bb96-50d059cb9b5b\") " Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.951455 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w25d8\" (UniqueName: \"kubernetes.io/projected/f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec-kube-api-access-w25d8\") pod \"openstack-operator-index-xqbxv\" (UID: \"f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec\") " pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.970549 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv" (OuterVolumeSpecName: "kube-api-access-z86gv") pod "cf896058-ab59-4f91-bb96-50d059cb9b5b" (UID: "cf896058-ab59-4f91-bb96-50d059cb9b5b"). InnerVolumeSpecName "kube-api-access-z86gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:28:05 crc kubenswrapper[4831]: I1124 08:28:05.976235 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w25d8\" (UniqueName: \"kubernetes.io/projected/f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec-kube-api-access-w25d8\") pod \"openstack-operator-index-xqbxv\" (UID: \"f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec\") " pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.044494 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.053258 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z86gv\" (UniqueName: \"kubernetes.io/projected/cf896058-ab59-4f91-bb96-50d059cb9b5b-kube-api-access-z86gv\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.469083 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xqbxv"] Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.565447 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xqbxv" event={"ID":"f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec","Type":"ContainerStarted","Data":"c76f8fafe2447819332bfa964801f2ea283a021497ed5f009720381d9b3a09a7"} Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.567518 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hxwpd" event={"ID":"cf896058-ab59-4f91-bb96-50d059cb9b5b","Type":"ContainerDied","Data":"97a66abcba53c74e270c89264f5955bcfaf8a1b58c1f1ad860664e747ec67126"} Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.567581 4831 scope.go:117] "RemoveContainer" containerID="e889962d07fe8db17224e20a58cd5ab6808c16e437ba8fc1aee880f8614e1147" Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.567743 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hxwpd" Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.616591 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.623885 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-hxwpd"] Nov 24 08:28:06 crc kubenswrapper[4831]: I1124 08:28:06.900526 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf896058-ab59-4f91-bb96-50d059cb9b5b" path="/var/lib/kubelet/pods/cf896058-ab59-4f91-bb96-50d059cb9b5b/volumes" Nov 24 08:28:07 crc kubenswrapper[4831]: I1124 08:28:07.577282 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xqbxv" event={"ID":"f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec","Type":"ContainerStarted","Data":"e9deda34822670d7dfea205ca9d63213ebe1a7dc39cfa77569b67162e1c5da0e"} Nov 24 08:28:07 crc kubenswrapper[4831]: I1124 08:28:07.594032 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xqbxv" podStartSLOduration=2.535404772 podStartE2EDuration="2.594016005s" podCreationTimestamp="2025-11-24 08:28:05 +0000 UTC" firstStartedPulling="2025-11-24 08:28:06.484660711 +0000 UTC m=+760.359805864" lastFinishedPulling="2025-11-24 08:28:06.543271904 +0000 UTC m=+760.418417097" observedRunningTime="2025-11-24 08:28:07.591889435 +0000 UTC m=+761.467034598" watchObservedRunningTime="2025-11-24 08:28:07.594016005 +0000 UTC m=+761.469161158" Nov 24 08:28:16 crc kubenswrapper[4831]: I1124 08:28:16.045389 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:16 crc kubenswrapper[4831]: I1124 08:28:16.046215 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:16 crc kubenswrapper[4831]: I1124 08:28:16.090451 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:16 crc kubenswrapper[4831]: I1124 08:28:16.683311 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xqbxv" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.534263 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:19 crc kubenswrapper[4831]: E1124 08:28:19.536839 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf896058-ab59-4f91-bb96-50d059cb9b5b" containerName="registry-server" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.536885 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf896058-ab59-4f91-bb96-50d059cb9b5b" containerName="registry-server" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.537007 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf896058-ab59-4f91-bb96-50d059cb9b5b" containerName="registry-server" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.537988 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.554638 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.692804 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.692970 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.693056 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkll6\" (UniqueName: \"kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.794394 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.794488 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkll6\" (UniqueName: \"kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.794618 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.795918 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.795982 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.828804 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkll6\" (UniqueName: \"kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6\") pod \"community-operators-n4dls\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:19 crc kubenswrapper[4831]: I1124 08:28:19.899247 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:20 crc kubenswrapper[4831]: I1124 08:28:20.319921 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:20 crc kubenswrapper[4831]: I1124 08:28:20.675188 4831 generic.go:334] "Generic (PLEG): container finished" podID="cc04b638-2326-4564-9921-e88bf9823c66" containerID="e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef" exitCode=0 Nov 24 08:28:20 crc kubenswrapper[4831]: I1124 08:28:20.675313 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerDied","Data":"e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef"} Nov 24 08:28:20 crc kubenswrapper[4831]: I1124 08:28:20.676604 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerStarted","Data":"6418807f036e0a814ed9a8b4cc55d119401bcad3c02d19584b1b87ecb274e80e"} Nov 24 08:28:21 crc kubenswrapper[4831]: I1124 08:28:21.687736 4831 generic.go:334] "Generic (PLEG): container finished" podID="cc04b638-2326-4564-9921-e88bf9823c66" containerID="ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d" exitCode=0 Nov 24 08:28:21 crc kubenswrapper[4831]: I1124 08:28:21.687802 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerDied","Data":"ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d"} Nov 24 08:28:22 crc kubenswrapper[4831]: I1124 08:28:22.696259 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerStarted","Data":"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081"} Nov 24 08:28:23 crc kubenswrapper[4831]: I1124 08:28:23.941575 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n4dls" podStartSLOduration=3.516797845 podStartE2EDuration="4.941553697s" podCreationTimestamp="2025-11-24 08:28:19 +0000 UTC" firstStartedPulling="2025-11-24 08:28:20.678277252 +0000 UTC m=+774.553422445" lastFinishedPulling="2025-11-24 08:28:22.103033134 +0000 UTC m=+775.978178297" observedRunningTime="2025-11-24 08:28:22.743391668 +0000 UTC m=+776.618536821" watchObservedRunningTime="2025-11-24 08:28:23.941553697 +0000 UTC m=+777.816698860" Nov 24 08:28:23 crc kubenswrapper[4831]: I1124 08:28:23.944425 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r"] Nov 24 08:28:23 crc kubenswrapper[4831]: I1124 08:28:23.945832 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:23 crc kubenswrapper[4831]: I1124 08:28:23.949166 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-27nsf" Nov 24 08:28:23 crc kubenswrapper[4831]: I1124 08:28:23.958212 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r"] Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.060555 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.060963 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnzhk\" (UniqueName: \"kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.061239 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.162465 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.162528 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnzhk\" (UniqueName: \"kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.162594 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.162969 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.163156 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.186388 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnzhk\" (UniqueName: \"kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk\") pod \"bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.266718 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.701908 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.704177 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.717193 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.777883 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r"] Nov 24 08:28:24 crc kubenswrapper[4831]: W1124 08:28:24.790117 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebab0eb6_34eb_480f_8d73_d30997857d21.slice/crio-ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d WatchSource:0}: Error finding container ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d: Status 404 returned error can't find the container with id ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.872439 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.872524 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.872549 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g979h\" (UniqueName: \"kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.973483 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.973567 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.973587 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g979h\" (UniqueName: \"kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.974044 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.974101 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:24 crc kubenswrapper[4831]: I1124 08:28:24.990986 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g979h\" (UniqueName: \"kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h\") pod \"redhat-marketplace-bgh49\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.028909 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.506705 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.721218 4831 generic.go:334] "Generic (PLEG): container finished" podID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerID="77b54deaf12f95a8d1a67c58830a6c1c6ea8ad7f6384cbca830629b10a796f54" exitCode=0 Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.721288 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" event={"ID":"ebab0eb6-34eb-480f-8d73-d30997857d21","Type":"ContainerDied","Data":"77b54deaf12f95a8d1a67c58830a6c1c6ea8ad7f6384cbca830629b10a796f54"} Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.721753 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" event={"ID":"ebab0eb6-34eb-480f-8d73-d30997857d21","Type":"ContainerStarted","Data":"ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d"} Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.723682 4831 generic.go:334] "Generic (PLEG): container finished" podID="82f9e0e6-531a-482b-8aae-048dcafde041" containerID="1f1a005b6f40307e83b607df7651c0a44ec97fcee080f7387bc121efa97ef40b" exitCode=0 Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.723775 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerDied","Data":"1f1a005b6f40307e83b607df7651c0a44ec97fcee080f7387bc121efa97ef40b"} Nov 24 08:28:25 crc kubenswrapper[4831]: I1124 08:28:25.723846 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerStarted","Data":"8d28bd084f57d6a0c905e83e9e2c540ed7793cf5e6416480589fece990bf53d2"} Nov 24 08:28:26 crc kubenswrapper[4831]: I1124 08:28:26.731697 4831 generic.go:334] "Generic (PLEG): container finished" podID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerID="a4fe91ff8284826b76c0d1369e6b4ec22b65281bd025236d4a9b47c8799a645f" exitCode=0 Nov 24 08:28:26 crc kubenswrapper[4831]: I1124 08:28:26.734382 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" event={"ID":"ebab0eb6-34eb-480f-8d73-d30997857d21","Type":"ContainerDied","Data":"a4fe91ff8284826b76c0d1369e6b4ec22b65281bd025236d4a9b47c8799a645f"} Nov 24 08:28:26 crc kubenswrapper[4831]: I1124 08:28:26.738183 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerStarted","Data":"a2b5dc81a8eb320873a5f9515389ed22d832557a39b35b6b1d9c01508d11d9e7"} Nov 24 08:28:27 crc kubenswrapper[4831]: E1124 08:28:27.074409 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebab0eb6_34eb_480f_8d73_d30997857d21.slice/crio-7375ea87b23cdb7e0e1cc3b548b300b11c72502f253160fcba98d48fd6b2b1ba.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:28:27 crc kubenswrapper[4831]: I1124 08:28:27.750345 4831 generic.go:334] "Generic (PLEG): container finished" podID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerID="7375ea87b23cdb7e0e1cc3b548b300b11c72502f253160fcba98d48fd6b2b1ba" exitCode=0 Nov 24 08:28:27 crc kubenswrapper[4831]: I1124 08:28:27.750433 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" event={"ID":"ebab0eb6-34eb-480f-8d73-d30997857d21","Type":"ContainerDied","Data":"7375ea87b23cdb7e0e1cc3b548b300b11c72502f253160fcba98d48fd6b2b1ba"} Nov 24 08:28:27 crc kubenswrapper[4831]: I1124 08:28:27.755625 4831 generic.go:334] "Generic (PLEG): container finished" podID="82f9e0e6-531a-482b-8aae-048dcafde041" containerID="a2b5dc81a8eb320873a5f9515389ed22d832557a39b35b6b1d9c01508d11d9e7" exitCode=0 Nov 24 08:28:27 crc kubenswrapper[4831]: I1124 08:28:27.755684 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerDied","Data":"a2b5dc81a8eb320873a5f9515389ed22d832557a39b35b6b1d9c01508d11d9e7"} Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.401724 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.402109 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.402161 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.402784 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.402836 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090" gracePeriod=600 Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.764011 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerStarted","Data":"b3881b615b566d61f0088efc4dbdf97b6ff81f913f2257ff4ffa2c5eae7d8e82"} Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.766353 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090" exitCode=0 Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.766381 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090"} Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.766429 4831 scope.go:117] "RemoveContainer" containerID="db9e0ef07b38faaeae940f42e5d89ffaa4ac6d82cbb0395ac1112b02f6e2bf36" Nov 24 08:28:28 crc kubenswrapper[4831]: I1124 08:28:28.789300 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bgh49" podStartSLOduration=2.34827753 podStartE2EDuration="4.789275158s" podCreationTimestamp="2025-11-24 08:28:24 +0000 UTC" firstStartedPulling="2025-11-24 08:28:25.726991243 +0000 UTC m=+779.602136396" lastFinishedPulling="2025-11-24 08:28:28.167988851 +0000 UTC m=+782.043134024" observedRunningTime="2025-11-24 08:28:28.785319856 +0000 UTC m=+782.660465029" watchObservedRunningTime="2025-11-24 08:28:28.789275158 +0000 UTC m=+782.664420351" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.172709 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.277839 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle\") pod \"ebab0eb6-34eb-480f-8d73-d30997857d21\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.277900 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util\") pod \"ebab0eb6-34eb-480f-8d73-d30997857d21\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.277951 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnzhk\" (UniqueName: \"kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk\") pod \"ebab0eb6-34eb-480f-8d73-d30997857d21\" (UID: \"ebab0eb6-34eb-480f-8d73-d30997857d21\") " Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.279144 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle" (OuterVolumeSpecName: "bundle") pod "ebab0eb6-34eb-480f-8d73-d30997857d21" (UID: "ebab0eb6-34eb-480f-8d73-d30997857d21"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.291746 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util" (OuterVolumeSpecName: "util") pod "ebab0eb6-34eb-480f-8d73-d30997857d21" (UID: "ebab0eb6-34eb-480f-8d73-d30997857d21"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.292983 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk" (OuterVolumeSpecName: "kube-api-access-hnzhk") pod "ebab0eb6-34eb-480f-8d73-d30997857d21" (UID: "ebab0eb6-34eb-480f-8d73-d30997857d21"). InnerVolumeSpecName "kube-api-access-hnzhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.379418 4831 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.379452 4831 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebab0eb6-34eb-480f-8d73-d30997857d21-util\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.379464 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnzhk\" (UniqueName: \"kubernetes.io/projected/ebab0eb6-34eb-480f-8d73-d30997857d21-kube-api-access-hnzhk\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.778770 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" event={"ID":"ebab0eb6-34eb-480f-8d73-d30997857d21","Type":"ContainerDied","Data":"ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d"} Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.779231 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ade4305dda66c81d097ca7f92be11b763445d88d34eac270022e1a99acf1a09d" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.778830 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.784198 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40"} Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.899933 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.899998 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:29 crc kubenswrapper[4831]: I1124 08:28:29.970030 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:30 crc kubenswrapper[4831]: I1124 08:28:30.864033 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:33 crc kubenswrapper[4831]: I1124 08:28:33.689641 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:33 crc kubenswrapper[4831]: I1124 08:28:33.690783 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n4dls" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="registry-server" containerID="cri-o://b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081" gracePeriod=2 Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.208423 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.351623 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities\") pod \"cc04b638-2326-4564-9921-e88bf9823c66\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.351686 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content\") pod \"cc04b638-2326-4564-9921-e88bf9823c66\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.351737 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkll6\" (UniqueName: \"kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6\") pod \"cc04b638-2326-4564-9921-e88bf9823c66\" (UID: \"cc04b638-2326-4564-9921-e88bf9823c66\") " Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.352577 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities" (OuterVolumeSpecName: "utilities") pod "cc04b638-2326-4564-9921-e88bf9823c66" (UID: "cc04b638-2326-4564-9921-e88bf9823c66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.371077 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6" (OuterVolumeSpecName: "kube-api-access-lkll6") pod "cc04b638-2326-4564-9921-e88bf9823c66" (UID: "cc04b638-2326-4564-9921-e88bf9823c66"). InnerVolumeSpecName "kube-api-access-lkll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.421439 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc04b638-2326-4564-9921-e88bf9823c66" (UID: "cc04b638-2326-4564-9921-e88bf9823c66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.453170 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.453203 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkll6\" (UniqueName: \"kubernetes.io/projected/cc04b638-2326-4564-9921-e88bf9823c66-kube-api-access-lkll6\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.453214 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc04b638-2326-4564-9921-e88bf9823c66-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.822227 4831 generic.go:334] "Generic (PLEG): container finished" podID="cc04b638-2326-4564-9921-e88bf9823c66" containerID="b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081" exitCode=0 Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.822673 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4dls" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.823165 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerDied","Data":"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081"} Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.823203 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4dls" event={"ID":"cc04b638-2326-4564-9921-e88bf9823c66","Type":"ContainerDied","Data":"6418807f036e0a814ed9a8b4cc55d119401bcad3c02d19584b1b87ecb274e80e"} Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.823243 4831 scope.go:117] "RemoveContainer" containerID="b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.838738 4831 scope.go:117] "RemoveContainer" containerID="ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.848963 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.867665 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n4dls"] Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.872336 4831 scope.go:117] "RemoveContainer" containerID="e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.891619 4831 scope.go:117] "RemoveContainer" containerID="b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081" Nov 24 08:28:34 crc kubenswrapper[4831]: E1124 08:28:34.892130 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081\": container with ID starting with b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081 not found: ID does not exist" containerID="b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.892162 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081"} err="failed to get container status \"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081\": rpc error: code = NotFound desc = could not find container \"b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081\": container with ID starting with b4a8ead6615d4f9e63fa0e8e1d96dff0a955ed4dba5cb315ae0835dfabc92081 not found: ID does not exist" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.892185 4831 scope.go:117] "RemoveContainer" containerID="ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d" Nov 24 08:28:34 crc kubenswrapper[4831]: E1124 08:28:34.892629 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d\": container with ID starting with ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d not found: ID does not exist" containerID="ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.892683 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d"} err="failed to get container status \"ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d\": rpc error: code = NotFound desc = could not find container \"ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d\": container with ID starting with ec2b4f564d44b0c676e8854b769b1d2d36f3503a347d91b543265490c0c29d3d not found: ID does not exist" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.892715 4831 scope.go:117] "RemoveContainer" containerID="e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef" Nov 24 08:28:34 crc kubenswrapper[4831]: E1124 08:28:34.893161 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef\": container with ID starting with e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef not found: ID does not exist" containerID="e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.893190 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef"} err="failed to get container status \"e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef\": rpc error: code = NotFound desc = could not find container \"e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef\": container with ID starting with e398c34df7172dda661a71aa4cbc15540473bf17f8ce3a366184e2e522cf5cef not found: ID does not exist" Nov 24 08:28:34 crc kubenswrapper[4831]: I1124 08:28:34.901873 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc04b638-2326-4564-9921-e88bf9823c66" path="/var/lib/kubelet/pods/cc04b638-2326-4564-9921-e88bf9823c66/volumes" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.029392 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.029445 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.067214 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.623710 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm"] Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624305 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="util" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624328 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="util" Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624339 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="extract-utilities" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624345 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="extract-utilities" Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624358 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="extract" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624365 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="extract" Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624373 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="pull" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624378 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="pull" Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624387 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="extract-content" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624394 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="extract-content" Nov 24 08:28:35 crc kubenswrapper[4831]: E1124 08:28:35.624404 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="registry-server" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624410 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="registry-server" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624531 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc04b638-2326-4564-9921-e88bf9823c66" containerName="registry-server" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.624543 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebab0eb6-34eb-480f-8d73-d30997857d21" containerName="extract" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.625115 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.632723 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-dtqn6" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.671260 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9hnz\" (UniqueName: \"kubernetes.io/projected/2940bb03-68aa-433e-a3c2-290fd7993dcd-kube-api-access-h9hnz\") pod \"openstack-operator-controller-operator-84f6799855-frnsm\" (UID: \"2940bb03-68aa-433e-a3c2-290fd7993dcd\") " pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.679343 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm"] Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.772732 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9hnz\" (UniqueName: \"kubernetes.io/projected/2940bb03-68aa-433e-a3c2-290fd7993dcd-kube-api-access-h9hnz\") pod \"openstack-operator-controller-operator-84f6799855-frnsm\" (UID: \"2940bb03-68aa-433e-a3c2-290fd7993dcd\") " pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.802519 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9hnz\" (UniqueName: \"kubernetes.io/projected/2940bb03-68aa-433e-a3c2-290fd7993dcd-kube-api-access-h9hnz\") pod \"openstack-operator-controller-operator-84f6799855-frnsm\" (UID: \"2940bb03-68aa-433e-a3c2-290fd7993dcd\") " pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.870816 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:35 crc kubenswrapper[4831]: I1124 08:28:35.942571 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:36 crc kubenswrapper[4831]: I1124 08:28:36.362016 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm"] Nov 24 08:28:36 crc kubenswrapper[4831]: W1124 08:28:36.365055 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2940bb03_68aa_433e_a3c2_290fd7993dcd.slice/crio-eb122730c17056dbb866967ec57ab1fd851df17269bb973904fe689eed892ab9 WatchSource:0}: Error finding container eb122730c17056dbb866967ec57ab1fd851df17269bb973904fe689eed892ab9: Status 404 returned error can't find the container with id eb122730c17056dbb866967ec57ab1fd851df17269bb973904fe689eed892ab9 Nov 24 08:28:36 crc kubenswrapper[4831]: I1124 08:28:36.842987 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" event={"ID":"2940bb03-68aa-433e-a3c2-290fd7993dcd","Type":"ContainerStarted","Data":"eb122730c17056dbb866967ec57ab1fd851df17269bb973904fe689eed892ab9"} Nov 24 08:28:37 crc kubenswrapper[4831]: I1124 08:28:37.888824 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:37 crc kubenswrapper[4831]: I1124 08:28:37.889464 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bgh49" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="registry-server" containerID="cri-o://b3881b615b566d61f0088efc4dbdf97b6ff81f913f2257ff4ffa2c5eae7d8e82" gracePeriod=2 Nov 24 08:28:38 crc kubenswrapper[4831]: I1124 08:28:38.863376 4831 generic.go:334] "Generic (PLEG): container finished" podID="82f9e0e6-531a-482b-8aae-048dcafde041" containerID="b3881b615b566d61f0088efc4dbdf97b6ff81f913f2257ff4ffa2c5eae7d8e82" exitCode=0 Nov 24 08:28:38 crc kubenswrapper[4831]: I1124 08:28:38.863770 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerDied","Data":"b3881b615b566d61f0088efc4dbdf97b6ff81f913f2257ff4ffa2c5eae7d8e82"} Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.694276 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.844474 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g979h\" (UniqueName: \"kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h\") pod \"82f9e0e6-531a-482b-8aae-048dcafde041\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.845039 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content\") pod \"82f9e0e6-531a-482b-8aae-048dcafde041\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.845072 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities\") pod \"82f9e0e6-531a-482b-8aae-048dcafde041\" (UID: \"82f9e0e6-531a-482b-8aae-048dcafde041\") " Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.850898 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities" (OuterVolumeSpecName: "utilities") pod "82f9e0e6-531a-482b-8aae-048dcafde041" (UID: "82f9e0e6-531a-482b-8aae-048dcafde041"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.852658 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h" (OuterVolumeSpecName: "kube-api-access-g979h") pod "82f9e0e6-531a-482b-8aae-048dcafde041" (UID: "82f9e0e6-531a-482b-8aae-048dcafde041"). InnerVolumeSpecName "kube-api-access-g979h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.863610 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82f9e0e6-531a-482b-8aae-048dcafde041" (UID: "82f9e0e6-531a-482b-8aae-048dcafde041"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.880580 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" event={"ID":"2940bb03-68aa-433e-a3c2-290fd7993dcd","Type":"ContainerStarted","Data":"a4be7734039c71ee444fc764f57d181327259c20f2200c2174ab025046d177fd"} Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.883474 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bgh49" event={"ID":"82f9e0e6-531a-482b-8aae-048dcafde041","Type":"ContainerDied","Data":"8d28bd084f57d6a0c905e83e9e2c540ed7793cf5e6416480589fece990bf53d2"} Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.883542 4831 scope.go:117] "RemoveContainer" containerID="b3881b615b566d61f0088efc4dbdf97b6ff81f913f2257ff4ffa2c5eae7d8e82" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.883596 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bgh49" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.907010 4831 scope.go:117] "RemoveContainer" containerID="a2b5dc81a8eb320873a5f9515389ed22d832557a39b35b6b1d9c01508d11d9e7" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.919481 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.924541 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bgh49"] Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.927751 4831 scope.go:117] "RemoveContainer" containerID="1f1a005b6f40307e83b607df7651c0a44ec97fcee080f7387bc121efa97ef40b" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.946533 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.946567 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f9e0e6-531a-482b-8aae-048dcafde041-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:40 crc kubenswrapper[4831]: I1124 08:28:40.946581 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g979h\" (UniqueName: \"kubernetes.io/projected/82f9e0e6-531a-482b-8aae-048dcafde041-kube-api-access-g979h\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:42 crc kubenswrapper[4831]: I1124 08:28:42.903898 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" path="/var/lib/kubelet/pods/82f9e0e6-531a-482b-8aae-048dcafde041/volumes" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.298578 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:43 crc kubenswrapper[4831]: E1124 08:28:43.298809 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="registry-server" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.298820 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="registry-server" Nov 24 08:28:43 crc kubenswrapper[4831]: E1124 08:28:43.298829 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="extract-utilities" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.298835 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="extract-utilities" Nov 24 08:28:43 crc kubenswrapper[4831]: E1124 08:28:43.298843 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="extract-content" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.298851 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="extract-content" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.298952 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f9e0e6-531a-482b-8aae-048dcafde041" containerName="registry-server" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.299720 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.308121 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.480504 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qzfm\" (UniqueName: \"kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.480573 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.480658 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.582691 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qzfm\" (UniqueName: \"kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.582772 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.582835 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.583465 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.584096 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.604050 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qzfm\" (UniqueName: \"kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm\") pod \"redhat-operators-rdkmh\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.619065 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.899203 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.918030 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" event={"ID":"2940bb03-68aa-433e-a3c2-290fd7993dcd","Type":"ContainerStarted","Data":"dd1297c692c452f9fbe72721f663cf8c5590bfb57f8530a605e5ac4758a9da41"} Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.919414 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:43 crc kubenswrapper[4831]: I1124 08:28:43.921402 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerStarted","Data":"861201eb98e1028adb5372d55f48b3f520f427ae0e40b0b83e6301013293a3fa"} Nov 24 08:28:44 crc kubenswrapper[4831]: I1124 08:28:44.931844 4831 generic.go:334] "Generic (PLEG): container finished" podID="f361072e-2a5f-4733-8597-73c4099121df" containerID="24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377" exitCode=0 Nov 24 08:28:44 crc kubenswrapper[4831]: I1124 08:28:44.931892 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerDied","Data":"24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377"} Nov 24 08:28:44 crc kubenswrapper[4831]: I1124 08:28:44.956553 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" podStartSLOduration=3.424469218 podStartE2EDuration="9.956528303s" podCreationTimestamp="2025-11-24 08:28:35 +0000 UTC" firstStartedPulling="2025-11-24 08:28:36.366926706 +0000 UTC m=+790.242071859" lastFinishedPulling="2025-11-24 08:28:42.898985791 +0000 UTC m=+796.774130944" observedRunningTime="2025-11-24 08:28:43.95320269 +0000 UTC m=+797.828347843" watchObservedRunningTime="2025-11-24 08:28:44.956528303 +0000 UTC m=+798.831673466" Nov 24 08:28:45 crc kubenswrapper[4831]: I1124 08:28:45.941689 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerStarted","Data":"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf"} Nov 24 08:28:45 crc kubenswrapper[4831]: I1124 08:28:45.944771 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-84f6799855-frnsm" Nov 24 08:28:46 crc kubenswrapper[4831]: I1124 08:28:46.949830 4831 generic.go:334] "Generic (PLEG): container finished" podID="f361072e-2a5f-4733-8597-73c4099121df" containerID="bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf" exitCode=0 Nov 24 08:28:46 crc kubenswrapper[4831]: I1124 08:28:46.950442 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerDied","Data":"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf"} Nov 24 08:28:47 crc kubenswrapper[4831]: I1124 08:28:47.958408 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerStarted","Data":"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914"} Nov 24 08:28:47 crc kubenswrapper[4831]: I1124 08:28:47.982899 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rdkmh" podStartSLOduration=2.373553797 podStartE2EDuration="4.982875984s" podCreationTimestamp="2025-11-24 08:28:43 +0000 UTC" firstStartedPulling="2025-11-24 08:28:44.933984557 +0000 UTC m=+798.809129700" lastFinishedPulling="2025-11-24 08:28:47.543306734 +0000 UTC m=+801.418451887" observedRunningTime="2025-11-24 08:28:47.976216426 +0000 UTC m=+801.851361589" watchObservedRunningTime="2025-11-24 08:28:47.982875984 +0000 UTC m=+801.858021137" Nov 24 08:28:53 crc kubenswrapper[4831]: I1124 08:28:53.619344 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:53 crc kubenswrapper[4831]: I1124 08:28:53.619683 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:53 crc kubenswrapper[4831]: I1124 08:28:53.675813 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:54 crc kubenswrapper[4831]: I1124 08:28:54.046653 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:54 crc kubenswrapper[4831]: I1124 08:28:54.097833 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.005994 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rdkmh" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="registry-server" containerID="cri-o://6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914" gracePeriod=2 Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.434025 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.565651 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content\") pod \"f361072e-2a5f-4733-8597-73c4099121df\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.565714 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qzfm\" (UniqueName: \"kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm\") pod \"f361072e-2a5f-4733-8597-73c4099121df\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.565784 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities\") pod \"f361072e-2a5f-4733-8597-73c4099121df\" (UID: \"f361072e-2a5f-4733-8597-73c4099121df\") " Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.567140 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities" (OuterVolumeSpecName: "utilities") pod "f361072e-2a5f-4733-8597-73c4099121df" (UID: "f361072e-2a5f-4733-8597-73c4099121df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.572077 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm" (OuterVolumeSpecName: "kube-api-access-5qzfm") pod "f361072e-2a5f-4733-8597-73c4099121df" (UID: "f361072e-2a5f-4733-8597-73c4099121df"). InnerVolumeSpecName "kube-api-access-5qzfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.652007 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f361072e-2a5f-4733-8597-73c4099121df" (UID: "f361072e-2a5f-4733-8597-73c4099121df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.667305 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.667353 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qzfm\" (UniqueName: \"kubernetes.io/projected/f361072e-2a5f-4733-8597-73c4099121df-kube-api-access-5qzfm\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:56 crc kubenswrapper[4831]: I1124 08:28:56.667365 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f361072e-2a5f-4733-8597-73c4099121df-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.016787 4831 generic.go:334] "Generic (PLEG): container finished" podID="f361072e-2a5f-4733-8597-73c4099121df" containerID="6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914" exitCode=0 Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.016857 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerDied","Data":"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914"} Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.016891 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rdkmh" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.016922 4831 scope.go:117] "RemoveContainer" containerID="6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.016902 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rdkmh" event={"ID":"f361072e-2a5f-4733-8597-73c4099121df","Type":"ContainerDied","Data":"861201eb98e1028adb5372d55f48b3f520f427ae0e40b0b83e6301013293a3fa"} Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.039515 4831 scope.go:117] "RemoveContainer" containerID="bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.046989 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.060807 4831 scope.go:117] "RemoveContainer" containerID="24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.062366 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rdkmh"] Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.090673 4831 scope.go:117] "RemoveContainer" containerID="6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914" Nov 24 08:28:57 crc kubenswrapper[4831]: E1124 08:28:57.091286 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914\": container with ID starting with 6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914 not found: ID does not exist" containerID="6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.091381 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914"} err="failed to get container status \"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914\": rpc error: code = NotFound desc = could not find container \"6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914\": container with ID starting with 6e41061e6ca2c79e02a28b279204f7eebddf3e7dbb5b4cde17df9dad35a4d914 not found: ID does not exist" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.091421 4831 scope.go:117] "RemoveContainer" containerID="bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf" Nov 24 08:28:57 crc kubenswrapper[4831]: E1124 08:28:57.091798 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf\": container with ID starting with bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf not found: ID does not exist" containerID="bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.091827 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf"} err="failed to get container status \"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf\": rpc error: code = NotFound desc = could not find container \"bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf\": container with ID starting with bf1d1406501b23a224c426cbc4e8d611bcc965f2dbb45a46eb11773c2c40e4cf not found: ID does not exist" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.091847 4831 scope.go:117] "RemoveContainer" containerID="24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377" Nov 24 08:28:57 crc kubenswrapper[4831]: E1124 08:28:57.092212 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377\": container with ID starting with 24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377 not found: ID does not exist" containerID="24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377" Nov 24 08:28:57 crc kubenswrapper[4831]: I1124 08:28:57.092242 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377"} err="failed to get container status \"24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377\": rpc error: code = NotFound desc = could not find container \"24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377\": container with ID starting with 24880e0c62825b4f866d84262671728ef5155025d50966d94009ad4442ae8377 not found: ID does not exist" Nov 24 08:28:58 crc kubenswrapper[4831]: I1124 08:28:58.901740 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f361072e-2a5f-4733-8597-73c4099121df" path="/var/lib/kubelet/pods/f361072e-2a5f-4733-8597-73c4099121df/volumes" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.887017 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk"] Nov 24 08:29:17 crc kubenswrapper[4831]: E1124 08:29:17.888024 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="registry-server" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.888042 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="registry-server" Nov 24 08:29:17 crc kubenswrapper[4831]: E1124 08:29:17.888062 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="extract-utilities" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.888070 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="extract-utilities" Nov 24 08:29:17 crc kubenswrapper[4831]: E1124 08:29:17.888081 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="extract-content" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.888091 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="extract-content" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.888223 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="f361072e-2a5f-4733-8597-73c4099121df" containerName="registry-server" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.888961 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.908972 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-fqs75" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.919513 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk"] Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.945392 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv"] Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.950609 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:17 crc kubenswrapper[4831]: I1124 08:29:17.974088 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-swrvd" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.021682 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.022772 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.029873 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wzqv9" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.044337 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.046818 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft52v\" (UniqueName: \"kubernetes.io/projected/156abc13-0fab-4e2c-a786-98ddf7c340b5-kube-api-access-ft52v\") pod \"cinder-operator-controller-manager-6498cbf48f-994pv\" (UID: \"156abc13-0fab-4e2c-a786-98ddf7c340b5\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.046901 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjv2r\" (UniqueName: \"kubernetes.io/projected/78bce3ee-0478-4a09-9568-9e62c29dc754-kube-api-access-fjv2r\") pod \"barbican-operator-controller-manager-75fb479bcc-x7xzk\" (UID: \"78bce3ee-0478-4a09-9568-9e62c29dc754\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.075449 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.128393 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.130036 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.135513 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-qd8lv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.150950 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.151648 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx29s\" (UniqueName: \"kubernetes.io/projected/cdc653a8-5c91-4da4-8489-380f2fd0b05d-kube-api-access-mx29s\") pod \"designate-operator-controller-manager-767ccfd65f-bdfgj\" (UID: \"cdc653a8-5c91-4da4-8489-380f2fd0b05d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.151719 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft52v\" (UniqueName: \"kubernetes.io/projected/156abc13-0fab-4e2c-a786-98ddf7c340b5-kube-api-access-ft52v\") pod \"cinder-operator-controller-manager-6498cbf48f-994pv\" (UID: \"156abc13-0fab-4e2c-a786-98ddf7c340b5\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.151780 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjv2r\" (UniqueName: \"kubernetes.io/projected/78bce3ee-0478-4a09-9568-9e62c29dc754-kube-api-access-fjv2r\") pod \"barbican-operator-controller-manager-75fb479bcc-x7xzk\" (UID: \"78bce3ee-0478-4a09-9568-9e62c29dc754\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.203822 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft52v\" (UniqueName: \"kubernetes.io/projected/156abc13-0fab-4e2c-a786-98ddf7c340b5-kube-api-access-ft52v\") pod \"cinder-operator-controller-manager-6498cbf48f-994pv\" (UID: \"156abc13-0fab-4e2c-a786-98ddf7c340b5\") " pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.207607 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.209040 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.215532 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-6dqzn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.251555 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.254267 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx29s\" (UniqueName: \"kubernetes.io/projected/cdc653a8-5c91-4da4-8489-380f2fd0b05d-kube-api-access-mx29s\") pod \"designate-operator-controller-manager-767ccfd65f-bdfgj\" (UID: \"cdc653a8-5c91-4da4-8489-380f2fd0b05d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.254390 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfbdg\" (UniqueName: \"kubernetes.io/projected/ed86dbad-6726-46f7-a11f-feb0647ab7c5-kube-api-access-kfbdg\") pod \"glance-operator-controller-manager-7969689c84-xcdjn\" (UID: \"ed86dbad-6726-46f7-a11f-feb0647ab7c5\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.255181 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkbfq\" (UniqueName: \"kubernetes.io/projected/af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc-kube-api-access-nkbfq\") pod \"heat-operator-controller-manager-56f54d6746-ppdhf\" (UID: \"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.263581 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.278041 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-ztxlz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.302417 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.305725 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjv2r\" (UniqueName: \"kubernetes.io/projected/78bce3ee-0478-4a09-9568-9e62c29dc754-kube-api-access-fjv2r\") pod \"barbican-operator-controller-manager-75fb479bcc-x7xzk\" (UID: \"78bce3ee-0478-4a09-9568-9e62c29dc754\") " pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.305885 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.310578 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.319049 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.336106 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.336299 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nbwps" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.336571 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.341457 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.342840 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.352433 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.355903 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkbfq\" (UniqueName: \"kubernetes.io/projected/af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc-kube-api-access-nkbfq\") pod \"heat-operator-controller-manager-56f54d6746-ppdhf\" (UID: \"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.356019 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2sqb\" (UniqueName: \"kubernetes.io/projected/054589e5-1de8-4a64-b5ad-2d1688cf8219-kube-api-access-t2sqb\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.356042 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.356075 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgtf8\" (UniqueName: \"kubernetes.io/projected/e94a7743-730d-4259-a584-59d1034cd654-kube-api-access-vgtf8\") pod \"horizon-operator-controller-manager-598f69df5d-rshdg\" (UID: \"e94a7743-730d-4259-a584-59d1034cd654\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.356100 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfbdg\" (UniqueName: \"kubernetes.io/projected/ed86dbad-6726-46f7-a11f-feb0647ab7c5-kube-api-access-kfbdg\") pod \"glance-operator-controller-manager-7969689c84-xcdjn\" (UID: \"ed86dbad-6726-46f7-a11f-feb0647ab7c5\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.359111 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-4chx4" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.361384 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.362570 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.370813 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-f9xlv" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.380823 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-zftw6"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.382358 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.392711 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx29s\" (UniqueName: \"kubernetes.io/projected/cdc653a8-5c91-4da4-8489-380f2fd0b05d-kube-api-access-mx29s\") pod \"designate-operator-controller-manager-767ccfd65f-bdfgj\" (UID: \"cdc653a8-5c91-4da4-8489-380f2fd0b05d\") " pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.394004 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-9jsvt" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.404994 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkbfq\" (UniqueName: \"kubernetes.io/projected/af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc-kube-api-access-nkbfq\") pod \"heat-operator-controller-manager-56f54d6746-ppdhf\" (UID: \"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc\") " pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.420881 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.436100 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfbdg\" (UniqueName: \"kubernetes.io/projected/ed86dbad-6726-46f7-a11f-feb0647ab7c5-kube-api-access-kfbdg\") pod \"glance-operator-controller-manager-7969689c84-xcdjn\" (UID: \"ed86dbad-6726-46f7-a11f-feb0647ab7c5\") " pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.439679 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.463309 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnjtf\" (UniqueName: \"kubernetes.io/projected/021fa54a-79c0-483a-96af-9f154175eb2f-kube-api-access-nnjtf\") pod \"manila-operator-controller-manager-58f887965d-zftw6\" (UID: \"021fa54a-79c0-483a-96af-9f154175eb2f\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.468646 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwq8z\" (UniqueName: \"kubernetes.io/projected/8bab5961-2a79-482a-8935-b3f84b6e5fbf-kube-api-access-rwq8z\") pod \"keystone-operator-controller-manager-7454b96578-zz4ns\" (UID: \"8bab5961-2a79-482a-8935-b3f84b6e5fbf\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.468758 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2sqb\" (UniqueName: \"kubernetes.io/projected/054589e5-1de8-4a64-b5ad-2d1688cf8219-kube-api-access-t2sqb\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.468799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.468831 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkw6\" (UniqueName: \"kubernetes.io/projected/4d9702a6-a116-4598-bc95-14c7023cacb0-kube-api-access-ffkw6\") pod \"ironic-operator-controller-manager-99b499f4-rrtwx\" (UID: \"4d9702a6-a116-4598-bc95-14c7023cacb0\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.473591 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgtf8\" (UniqueName: \"kubernetes.io/projected/e94a7743-730d-4259-a584-59d1034cd654-kube-api-access-vgtf8\") pod \"horizon-operator-controller-manager-598f69df5d-rshdg\" (UID: \"e94a7743-730d-4259-a584-59d1034cd654\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:18 crc kubenswrapper[4831]: E1124 08:29:18.474431 4831 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:18 crc kubenswrapper[4831]: E1124 08:29:18.474585 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert podName:054589e5-1de8-4a64-b5ad-2d1688cf8219 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:18.974534067 +0000 UTC m=+832.849679370 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert") pod "infra-operator-controller-manager-6dd8864d7c-m5tlh" (UID: "054589e5-1de8-4a64-b5ad-2d1688cf8219") : secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.492946 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.496068 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.498288 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.505739 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.509109 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-zd5gb" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.518438 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgtf8\" (UniqueName: \"kubernetes.io/projected/e94a7743-730d-4259-a584-59d1034cd654-kube-api-access-vgtf8\") pod \"horizon-operator-controller-manager-598f69df5d-rshdg\" (UID: \"e94a7743-730d-4259-a584-59d1034cd654\") " pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.521290 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2sqb\" (UniqueName: \"kubernetes.io/projected/054589e5-1de8-4a64-b5ad-2d1688cf8219-kube-api-access-t2sqb\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.529751 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.530988 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.552134 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-zftw6"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.552205 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.553648 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.554949 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.555172 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cmnwx" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.576218 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnjtf\" (UniqueName: \"kubernetes.io/projected/021fa54a-79c0-483a-96af-9f154175eb2f-kube-api-access-nnjtf\") pod \"manila-operator-controller-manager-58f887965d-zftw6\" (UID: \"021fa54a-79c0-483a-96af-9f154175eb2f\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.576288 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwq8z\" (UniqueName: \"kubernetes.io/projected/8bab5961-2a79-482a-8935-b3f84b6e5fbf-kube-api-access-rwq8z\") pod \"keystone-operator-controller-manager-7454b96578-zz4ns\" (UID: \"8bab5961-2a79-482a-8935-b3f84b6e5fbf\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.576347 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szddl\" (UniqueName: \"kubernetes.io/projected/37a35a63-0850-4615-afe3-905c1218bbcd-kube-api-access-szddl\") pod \"mariadb-operator-controller-manager-54b5986bb8-zp88s\" (UID: \"37a35a63-0850-4615-afe3-905c1218bbcd\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.576421 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkw6\" (UniqueName: \"kubernetes.io/projected/4d9702a6-a116-4598-bc95-14c7023cacb0-kube-api-access-ffkw6\") pod \"ironic-operator-controller-manager-99b499f4-rrtwx\" (UID: \"4d9702a6-a116-4598-bc95-14c7023cacb0\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.576500 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql4ct\" (UniqueName: \"kubernetes.io/projected/e14d1d45-f2ee-4f9d-8260-e10bf3c8333c-kube-api-access-ql4ct\") pod \"neutron-operator-controller-manager-78bd47f458-hv6f6\" (UID: \"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.584653 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.584985 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-dk8ks" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.608044 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.621406 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.664397 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.666756 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwq8z\" (UniqueName: \"kubernetes.io/projected/8bab5961-2a79-482a-8935-b3f84b6e5fbf-kube-api-access-rwq8z\") pod \"keystone-operator-controller-manager-7454b96578-zz4ns\" (UID: \"8bab5961-2a79-482a-8935-b3f84b6e5fbf\") " pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.668753 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.669688 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnjtf\" (UniqueName: \"kubernetes.io/projected/021fa54a-79c0-483a-96af-9f154175eb2f-kube-api-access-nnjtf\") pod \"manila-operator-controller-manager-58f887965d-zftw6\" (UID: \"021fa54a-79c0-483a-96af-9f154175eb2f\") " pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.685381 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sf7k\" (UniqueName: \"kubernetes.io/projected/4bc8600a-62b5-4733-90cc-b08b66b605fa-kube-api-access-9sf7k\") pod \"nova-operator-controller-manager-cfbb9c588-n45zz\" (UID: \"4bc8600a-62b5-4733-90cc-b08b66b605fa\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.686830 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql4ct\" (UniqueName: \"kubernetes.io/projected/e14d1d45-f2ee-4f9d-8260-e10bf3c8333c-kube-api-access-ql4ct\") pod \"neutron-operator-controller-manager-78bd47f458-hv6f6\" (UID: \"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.687091 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szddl\" (UniqueName: \"kubernetes.io/projected/37a35a63-0850-4615-afe3-905c1218bbcd-kube-api-access-szddl\") pod \"mariadb-operator-controller-manager-54b5986bb8-zp88s\" (UID: \"37a35a63-0850-4615-afe3-905c1218bbcd\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.689660 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.743561 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.758040 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-94m4b" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.771055 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.789820 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dqns\" (UniqueName: \"kubernetes.io/projected/14cc197a-4c42-46dc-89c6-941dd1997aa8-kube-api-access-4dqns\") pod \"octavia-operator-controller-manager-54cfbf4c7d-n8nkt\" (UID: \"14cc197a-4c42-46dc-89c6-941dd1997aa8\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.789915 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sf7k\" (UniqueName: \"kubernetes.io/projected/4bc8600a-62b5-4733-90cc-b08b66b605fa-kube-api-access-9sf7k\") pod \"nova-operator-controller-manager-cfbb9c588-n45zz\" (UID: \"4bc8600a-62b5-4733-90cc-b08b66b605fa\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.823836 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.824654 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.828873 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szddl\" (UniqueName: \"kubernetes.io/projected/37a35a63-0850-4615-afe3-905c1218bbcd-kube-api-access-szddl\") pod \"mariadb-operator-controller-manager-54b5986bb8-zp88s\" (UID: \"37a35a63-0850-4615-afe3-905c1218bbcd\") " pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.832173 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql4ct\" (UniqueName: \"kubernetes.io/projected/e14d1d45-f2ee-4f9d-8260-e10bf3c8333c-kube-api-access-ql4ct\") pod \"neutron-operator-controller-manager-78bd47f458-hv6f6\" (UID: \"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c\") " pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.858736 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sf7k\" (UniqueName: \"kubernetes.io/projected/4bc8600a-62b5-4733-90cc-b08b66b605fa-kube-api-access-9sf7k\") pod \"nova-operator-controller-manager-cfbb9c588-n45zz\" (UID: \"4bc8600a-62b5-4733-90cc-b08b66b605fa\") " pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.895584 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dqns\" (UniqueName: \"kubernetes.io/projected/14cc197a-4c42-46dc-89c6-941dd1997aa8-kube-api-access-4dqns\") pod \"octavia-operator-controller-manager-54cfbf4c7d-n8nkt\" (UID: \"14cc197a-4c42-46dc-89c6-941dd1997aa8\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.899494 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.931675 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.964499 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.965522 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.968257 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dqns\" (UniqueName: \"kubernetes.io/projected/14cc197a-4c42-46dc-89c6-941dd1997aa8-kube-api-access-4dqns\") pod \"octavia-operator-controller-manager-54cfbf4c7d-n8nkt\" (UID: \"14cc197a-4c42-46dc-89c6-941dd1997aa8\") " pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.968780 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.985294 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n"] Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.990182 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-xrt8b" Nov 24 08:29:18 crc kubenswrapper[4831]: I1124 08:29:18.990390 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.007184 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.007431 4831 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.007482 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert podName:054589e5-1de8-4a64-b5ad-2d1688cf8219 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:20.0074651 +0000 UTC m=+833.882610253 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert") pod "infra-operator-controller-manager-6dd8864d7c-m5tlh" (UID: "054589e5-1de8-4a64-b5ad-2d1688cf8219") : secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.015330 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.016506 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.024652 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-cgtbl" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.034249 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkw6\" (UniqueName: \"kubernetes.io/projected/4d9702a6-a116-4598-bc95-14c7023cacb0-kube-api-access-ffkw6\") pod \"ironic-operator-controller-manager-99b499f4-rrtwx\" (UID: \"4d9702a6-a116-4598-bc95-14c7023cacb0\") " pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.050170 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.058486 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.073703 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-85c8r"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.075117 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.078822 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-l55rl" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.087624 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.100392 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.102215 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.108809 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-q26f7" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.110152 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sstg\" (UniqueName: \"kubernetes.io/projected/7531434a-d014-4bcb-9765-dd8ca784de62-kube-api-access-4sstg\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.110249 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.110284 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lcnw\" (UniqueName: \"kubernetes.io/projected/907add29-1677-495e-b968-432dd37c5e4e-kube-api-access-5lcnw\") pod \"ovn-operator-controller-manager-54fc5f65b7-4c6fw\" (UID: \"907add29-1677-495e-b968-432dd37c5e4e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.113597 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-85c8r"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.117944 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.139498 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.140901 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.154925 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-c6xv5" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.181365 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.203902 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9d55q"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.205149 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.207747 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.212802 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.216851 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-btdg2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.217110 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9d55q"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.217851 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkxx\" (UniqueName: \"kubernetes.io/projected/863fe05e-d2e1-47be-b0ab-1d69c2455dc1-kube-api-access-rjkxx\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4s2q2\" (UID: \"863fe05e-d2e1-47be-b0ab-1d69c2455dc1\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.217959 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sstg\" (UniqueName: \"kubernetes.io/projected/7531434a-d014-4bcb-9765-dd8ca784de62-kube-api-access-4sstg\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.218055 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgvlk\" (UniqueName: \"kubernetes.io/projected/e7c48858-2602-4763-b69e-96a4f7db47f5-kube-api-access-fgvlk\") pod \"placement-operator-controller-manager-5b797b8dff-md74d\" (UID: \"e7c48858-2602-4763-b69e-96a4f7db47f5\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.218143 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2svxh\" (UniqueName: \"kubernetes.io/projected/905703f5-8abe-45db-8e0e-a01369ef3413-kube-api-access-2svxh\") pod \"swift-operator-controller-manager-d656998f4-85c8r\" (UID: \"905703f5-8abe-45db-8e0e-a01369ef3413\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.218243 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.218348 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lcnw\" (UniqueName: \"kubernetes.io/projected/907add29-1677-495e-b968-432dd37c5e4e-kube-api-access-5lcnw\") pod \"ovn-operator-controller-manager-54fc5f65b7-4c6fw\" (UID: \"907add29-1677-495e-b968-432dd37c5e4e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.218924 4831 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.219034 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert podName:7531434a-d014-4bcb-9765-dd8ca784de62 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:19.719020148 +0000 UTC m=+833.594165301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" (UID: "7531434a-d014-4bcb-9765-dd8ca784de62") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.219497 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-vwkfz" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.220469 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.297050 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sstg\" (UniqueName: \"kubernetes.io/projected/7531434a-d014-4bcb-9765-dd8ca784de62-kube-api-access-4sstg\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.304786 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lcnw\" (UniqueName: \"kubernetes.io/projected/907add29-1677-495e-b968-432dd37c5e4e-kube-api-access-5lcnw\") pod \"ovn-operator-controller-manager-54fc5f65b7-4c6fw\" (UID: \"907add29-1677-495e-b968-432dd37c5e4e\") " pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.320189 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6mbr\" (UniqueName: \"kubernetes.io/projected/03d64575-b4f3-4b39-9fb8-d9597cea126c-kube-api-access-q6mbr\") pod \"watcher-operator-controller-manager-8c6448b9f-k5tgn\" (UID: \"03d64575-b4f3-4b39-9fb8-d9597cea126c\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.320420 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l862h\" (UniqueName: \"kubernetes.io/projected/0ee539a7-3c8d-47b9-bf94-bcecc7ff9333-kube-api-access-l862h\") pod \"test-operator-controller-manager-b4c496f69-9d55q\" (UID: \"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.320546 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkxx\" (UniqueName: \"kubernetes.io/projected/863fe05e-d2e1-47be-b0ab-1d69c2455dc1-kube-api-access-rjkxx\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4s2q2\" (UID: \"863fe05e-d2e1-47be-b0ab-1d69c2455dc1\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.320659 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgvlk\" (UniqueName: \"kubernetes.io/projected/e7c48858-2602-4763-b69e-96a4f7db47f5-kube-api-access-fgvlk\") pod \"placement-operator-controller-manager-5b797b8dff-md74d\" (UID: \"e7c48858-2602-4763-b69e-96a4f7db47f5\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.320799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2svxh\" (UniqueName: \"kubernetes.io/projected/905703f5-8abe-45db-8e0e-a01369ef3413-kube-api-access-2svxh\") pod \"swift-operator-controller-manager-d656998f4-85c8r\" (UID: \"905703f5-8abe-45db-8e0e-a01369ef3413\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.361311 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.362297 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.381470 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.381858 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.399555 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-78h5c" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.425378 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pb6n\" (UniqueName: \"kubernetes.io/projected/145ad046-da71-432f-9b5d-0f2c599c85e0-kube-api-access-4pb6n\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.425433 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6mbr\" (UniqueName: \"kubernetes.io/projected/03d64575-b4f3-4b39-9fb8-d9597cea126c-kube-api-access-q6mbr\") pod \"watcher-operator-controller-manager-8c6448b9f-k5tgn\" (UID: \"03d64575-b4f3-4b39-9fb8-d9597cea126c\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.425465 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l862h\" (UniqueName: \"kubernetes.io/projected/0ee539a7-3c8d-47b9-bf94-bcecc7ff9333-kube-api-access-l862h\") pod \"test-operator-controller-manager-b4c496f69-9d55q\" (UID: \"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.425491 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.434345 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkxx\" (UniqueName: \"kubernetes.io/projected/863fe05e-d2e1-47be-b0ab-1d69c2455dc1-kube-api-access-rjkxx\") pod \"telemetry-operator-controller-manager-6d4bf84b58-4s2q2\" (UID: \"863fe05e-d2e1-47be-b0ab-1d69c2455dc1\") " pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.494960 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgvlk\" (UniqueName: \"kubernetes.io/projected/e7c48858-2602-4763-b69e-96a4f7db47f5-kube-api-access-fgvlk\") pod \"placement-operator-controller-manager-5b797b8dff-md74d\" (UID: \"e7c48858-2602-4763-b69e-96a4f7db47f5\") " pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.510612 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2svxh\" (UniqueName: \"kubernetes.io/projected/905703f5-8abe-45db-8e0e-a01369ef3413-kube-api-access-2svxh\") pod \"swift-operator-controller-manager-d656998f4-85c8r\" (UID: \"905703f5-8abe-45db-8e0e-a01369ef3413\") " pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.542974 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6mbr\" (UniqueName: \"kubernetes.io/projected/03d64575-b4f3-4b39-9fb8-d9597cea126c-kube-api-access-q6mbr\") pod \"watcher-operator-controller-manager-8c6448b9f-k5tgn\" (UID: \"03d64575-b4f3-4b39-9fb8-d9597cea126c\") " pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.543875 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pb6n\" (UniqueName: \"kubernetes.io/projected/145ad046-da71-432f-9b5d-0f2c599c85e0-kube-api-access-4pb6n\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.544830 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.545110 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.546151 4831 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.546189 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert podName:145ad046-da71-432f-9b5d-0f2c599c85e0 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:20.046177237 +0000 UTC m=+833.921322390 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert") pod "openstack-operator-controller-manager-5dbdc8d64b-pcjw6" (UID: "145ad046-da71-432f-9b5d-0f2c599c85e0") : secret "webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.582542 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l862h\" (UniqueName: \"kubernetes.io/projected/0ee539a7-3c8d-47b9-bf94-bcecc7ff9333-kube-api-access-l862h\") pod \"test-operator-controller-manager-b4c496f69-9d55q\" (UID: \"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333\") " pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.608311 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pb6n\" (UniqueName: \"kubernetes.io/projected/145ad046-da71-432f-9b5d-0f2c599c85e0-kube-api-access-4pb6n\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.700803 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.717704 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.725932 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.748706 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.749062 4831 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: E1124 08:29:19.749188 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert podName:7531434a-d014-4bcb-9765-dd8ca784de62 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:20.749164593 +0000 UTC m=+834.624309746 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert") pod "openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" (UID: "7531434a-d014-4bcb-9765-dd8ca784de62") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.758687 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.782636 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.784679 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.809725 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.818190 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr"] Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.818246 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-qd7mz" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.849838 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qql4w\" (UniqueName: \"kubernetes.io/projected/82cf04aa-e08a-4626-ba4c-8c2321aaeb13-kube-api-access-qql4w\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-52vxr\" (UID: \"82cf04aa-e08a-4626-ba4c-8c2321aaeb13\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" Nov 24 08:29:19 crc kubenswrapper[4831]: I1124 08:29:19.951799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qql4w\" (UniqueName: \"kubernetes.io/projected/82cf04aa-e08a-4626-ba4c-8c2321aaeb13-kube-api-access-qql4w\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-52vxr\" (UID: \"82cf04aa-e08a-4626-ba4c-8c2321aaeb13\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.027526 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qql4w\" (UniqueName: \"kubernetes.io/projected/82cf04aa-e08a-4626-ba4c-8c2321aaeb13-kube-api-access-qql4w\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-52vxr\" (UID: \"82cf04aa-e08a-4626-ba4c-8c2321aaeb13\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.054092 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.054167 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:20 crc kubenswrapper[4831]: E1124 08:29:20.054409 4831 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 24 08:29:20 crc kubenswrapper[4831]: E1124 08:29:20.054474 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert podName:145ad046-da71-432f-9b5d-0f2c599c85e0 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:21.054443135 +0000 UTC m=+834.929588288 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert") pod "openstack-operator-controller-manager-5dbdc8d64b-pcjw6" (UID: "145ad046-da71-432f-9b5d-0f2c599c85e0") : secret "webhook-server-cert" not found Nov 24 08:29:20 crc kubenswrapper[4831]: E1124 08:29:20.054903 4831 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:20 crc kubenswrapper[4831]: E1124 08:29:20.054958 4831 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert podName:054589e5-1de8-4a64-b5ad-2d1688cf8219 nodeName:}" failed. No retries permitted until 2025-11-24 08:29:22.05495086 +0000 UTC m=+835.930096013 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert") pod "infra-operator-controller-manager-6dd8864d7c-m5tlh" (UID: "054589e5-1de8-4a64-b5ad-2d1688cf8219") : secret "infra-operator-webhook-server-cert" not found Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.159752 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.679149 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv"] Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.702454 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn"] Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.720218 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns"] Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.767658 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.790910 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7531434a-d014-4bcb-9765-dd8ca784de62-cert\") pod \"openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n\" (UID: \"7531434a-d014-4bcb-9765-dd8ca784de62\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:20 crc kubenswrapper[4831]: I1124 08:29:20.817609 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.074771 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.082926 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/145ad046-da71-432f-9b5d-0f2c599c85e0-cert\") pod \"openstack-operator-controller-manager-5dbdc8d64b-pcjw6\" (UID: \"145ad046-da71-432f-9b5d-0f2c599c85e0\") " pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.224241 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.262208 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" event={"ID":"ed86dbad-6726-46f7-a11f-feb0647ab7c5","Type":"ContainerStarted","Data":"0eda4a743d358e02bfd293f4ac0c43276404637f82cf021df20899884d29026d"} Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.265197 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.277646 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" event={"ID":"156abc13-0fab-4e2c-a786-98ddf7c340b5","Type":"ContainerStarted","Data":"928757111d9325171091580fa0ba6b6283ef0157737f0537e3fd100d48cd2f7a"} Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.290856 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" event={"ID":"8bab5961-2a79-482a-8935-b3f84b6e5fbf","Type":"ContainerStarted","Data":"c1d1037a9f35c9eb70711e3e00828d80bc251b4a66a7240c47a0ff0af2d2dce7"} Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.299389 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58f887965d-zftw6"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.328944 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.338579 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.350727 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.353778 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.366426 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.387474 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.402989 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg"] Nov 24 08:29:21 crc kubenswrapper[4831]: W1124 08:29:21.478709 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod907add29_1677_495e_b968_432dd37c5e4e.slice/crio-5972b8e232a30e43fd27f8a1197c9147bbee80944801ec0aa60c8928afb71395 WatchSource:0}: Error finding container 5972b8e232a30e43fd27f8a1197c9147bbee80944801ec0aa60c8928afb71395: Status 404 returned error can't find the container with id 5972b8e232a30e43fd27f8a1197c9147bbee80944801ec0aa60c8928afb71395 Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.613384 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d"] Nov 24 08:29:21 crc kubenswrapper[4831]: W1124 08:29:21.656512 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7c48858_2602_4763_b69e_96a4f7db47f5.slice/crio-d9b769394fad656f10443683396c8863dab55d6dc4502ddfa341d79e630f5a86 WatchSource:0}: Error finding container d9b769394fad656f10443683396c8863dab55d6dc4502ddfa341d79e630f5a86: Status 404 returned error can't find the container with id d9b769394fad656f10443683396c8863dab55d6dc4502ddfa341d79e630f5a86 Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.709209 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-d656998f4-85c8r"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.720549 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-b4c496f69-9d55q"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.747973 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.768469 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2"] Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.798640 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn"] Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.834123 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rjkxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6d4bf84b58-4s2q2_openstack-operators(863fe05e-d2e1-47be-b0ab-1d69c2455dc1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.851639 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n"] Nov 24 08:29:21 crc kubenswrapper[4831]: W1124 08:29:21.854544 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82cf04aa_e08a_4626_ba4c_8c2321aaeb13.slice/crio-122d2fb220c79b072d0c57ae03dac1fb637dcc6e8f123fa4391d9ef750eda389 WatchSource:0}: Error finding container 122d2fb220c79b072d0c57ae03dac1fb637dcc6e8f123fa4391d9ef750eda389: Status 404 returned error can't find the container with id 122d2fb220c79b072d0c57ae03dac1fb637dcc6e8f123fa4391d9ef750eda389 Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.886870 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx"] Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.894442 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qql4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-52vxr_openstack-operators(82cf04aa-e08a-4626-ba4c-8c2321aaeb13): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.896298 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podUID="82cf04aa-e08a-4626-ba4c-8c2321aaeb13" Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.931611 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4dqns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-54cfbf4c7d-n8nkt_openstack-operators(14cc197a-4c42-46dc-89c6-941dd1997aa8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.952176 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6mbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-k5tgn_openstack-operators(03d64575-b4f3-4b39-9fb8-d9597cea126c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 08:29:21 crc kubenswrapper[4831]: E1124 08:29:21.962138 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4sstg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n_openstack-operators(7531434a-d014-4bcb-9765-dd8ca784de62): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 24 08:29:21 crc kubenswrapper[4831]: I1124 08:29:21.964110 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt"] Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.108504 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.118086 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/054589e5-1de8-4a64-b5ad-2d1688cf8219-cert\") pod \"infra-operator-controller-manager-6dd8864d7c-m5tlh\" (UID: \"054589e5-1de8-4a64-b5ad-2d1688cf8219\") " pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.155365 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6"] Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.343662 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.390713 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" event={"ID":"4d9702a6-a116-4598-bc95-14c7023cacb0","Type":"ContainerStarted","Data":"5d8b89bfc485c48f43c6fde2c22b730558c0d77d68f7dda8095ef49b7a6974bb"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.412733 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" event={"ID":"78bce3ee-0478-4a09-9568-9e62c29dc754","Type":"ContainerStarted","Data":"47dad69fcf46b2478db38635bdac4396a3cf0eab7e8666c3cdf9d25203ed5112"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.459141 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" event={"ID":"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc","Type":"ContainerStarted","Data":"18a79a9cf645248c18637d0d47066e6eddfb0ffb55c26206c5e20e9a9e8f3d10"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.518172 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" event={"ID":"905703f5-8abe-45db-8e0e-a01369ef3413","Type":"ContainerStarted","Data":"f0fc2c070442871eeef2f668cc7d9b4e24ab73e3179219c4911f7f1ee641185d"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.520372 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" event={"ID":"14cc197a-4c42-46dc-89c6-941dd1997aa8","Type":"ContainerStarted","Data":"6061f7b996955e66f1ff9cde89dfd1f8f53a163770240d76010ef74f509c62ca"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.533119 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" event={"ID":"82cf04aa-e08a-4626-ba4c-8c2321aaeb13","Type":"ContainerStarted","Data":"122d2fb220c79b072d0c57ae03dac1fb637dcc6e8f123fa4391d9ef750eda389"} Nov 24 08:29:22 crc kubenswrapper[4831]: E1124 08:29:22.544780 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podUID="82cf04aa-e08a-4626-ba4c-8c2321aaeb13" Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.563434 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" event={"ID":"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c","Type":"ContainerStarted","Data":"1ac92da3d597d8bf0a1479812078527a33032bcca82a71396c03222bd8d24688"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.575211 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" event={"ID":"4bc8600a-62b5-4733-90cc-b08b66b605fa","Type":"ContainerStarted","Data":"0e0eba397660128182f80d3ddfdb7429dad8284aafad626c5b0eba730f7482d3"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.577282 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" event={"ID":"37a35a63-0850-4615-afe3-905c1218bbcd","Type":"ContainerStarted","Data":"c946127055e3b789b3f320acacc3cab304eb4a130c878c702452d3efc5239bdb"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.578837 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" event={"ID":"863fe05e-d2e1-47be-b0ab-1d69c2455dc1","Type":"ContainerStarted","Data":"42a13d0345879181b2aec956c6f98f1be6e0d2207cb284de45a5d782154439f6"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.579777 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" event={"ID":"cdc653a8-5c91-4da4-8489-380f2fd0b05d","Type":"ContainerStarted","Data":"bb9eeb7eec638a290a39f4a1b096c557d3613ebc26d30d0fe56b5490c1f4ac3d"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.581759 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" event={"ID":"e94a7743-730d-4259-a584-59d1034cd654","Type":"ContainerStarted","Data":"308ec50cad1b5e5799b08d4a018b43fe4ddc08b98a39aa1640fbb0aa47ed23f8"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.582717 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" event={"ID":"7531434a-d014-4bcb-9765-dd8ca784de62","Type":"ContainerStarted","Data":"5322bf1cd63411d8814b47d0fb3f42792856713f00a8a21ceb54dfab15d47063"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.583878 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" event={"ID":"021fa54a-79c0-483a-96af-9f154175eb2f","Type":"ContainerStarted","Data":"3f67357261923fb678e2fe6e135f52feb8ae3093a4d5f780d9b80c72d45da06d"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.584807 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" event={"ID":"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333","Type":"ContainerStarted","Data":"de772d84aa4d4dab4e1124697f71a320efcaec8756c08b0c102c023699d2eb97"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.585763 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" event={"ID":"03d64575-b4f3-4b39-9fb8-d9597cea126c","Type":"ContainerStarted","Data":"ac664c259bb80feb14b485b2e243d031ce3e7245496bf19d43e4dc9e257d0320"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.587013 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" event={"ID":"907add29-1677-495e-b968-432dd37c5e4e","Type":"ContainerStarted","Data":"5972b8e232a30e43fd27f8a1197c9147bbee80944801ec0aa60c8928afb71395"} Nov 24 08:29:22 crc kubenswrapper[4831]: I1124 08:29:22.587974 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" event={"ID":"e7c48858-2602-4763-b69e-96a4f7db47f5","Type":"ContainerStarted","Data":"d9b769394fad656f10443683396c8863dab55d6dc4502ddfa341d79e630f5a86"} Nov 24 08:29:22 crc kubenswrapper[4831]: E1124 08:29:22.627876 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" podUID="863fe05e-d2e1-47be-b0ab-1d69c2455dc1" Nov 24 08:29:22 crc kubenswrapper[4831]: E1124 08:29:22.799574 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podUID="03d64575-b4f3-4b39-9fb8-d9597cea126c" Nov 24 08:29:22 crc kubenswrapper[4831]: E1124 08:29:22.900871 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" podUID="14cc197a-4c42-46dc-89c6-941dd1997aa8" Nov 24 08:29:22 crc kubenswrapper[4831]: E1124 08:29:22.920347 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" podUID="7531434a-d014-4bcb-9765-dd8ca784de62" Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.317238 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh"] Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.618932 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" event={"ID":"14cc197a-4c42-46dc-89c6-941dd1997aa8","Type":"ContainerStarted","Data":"4be8c2f59ff4381c336af977f83afa526b62a29f4b88c51c902749b90f4bae48"} Nov 24 08:29:23 crc kubenswrapper[4831]: E1124 08:29:23.621433 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" podUID="14cc197a-4c42-46dc-89c6-941dd1997aa8" Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.628088 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" event={"ID":"145ad046-da71-432f-9b5d-0f2c599c85e0","Type":"ContainerStarted","Data":"8e59e1f9704b4b15d9417cd2099a7dccf498d03d6c922ea249abd5f363dbebb5"} Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.628145 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" event={"ID":"145ad046-da71-432f-9b5d-0f2c599c85e0","Type":"ContainerStarted","Data":"6bb0a2cc343b249cf05bd4075f6073c64890773be6e07badae8413ce254ff4f3"} Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.649774 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" event={"ID":"7531434a-d014-4bcb-9765-dd8ca784de62","Type":"ContainerStarted","Data":"9690565a955a047a098e40d11178ba62dfafb35b5e5d8d37661f2608cd66a6ef"} Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.653222 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" event={"ID":"054589e5-1de8-4a64-b5ad-2d1688cf8219","Type":"ContainerStarted","Data":"a894da46c779ffc8f850f6818437a8f2938ef982df51c938af82ea43957e91cb"} Nov 24 08:29:23 crc kubenswrapper[4831]: E1124 08:29:23.657754 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" podUID="7531434a-d014-4bcb-9765-dd8ca784de62" Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.692650 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" event={"ID":"03d64575-b4f3-4b39-9fb8-d9597cea126c","Type":"ContainerStarted","Data":"d79db519e15500e3ad3ae540ce436808a99f9cbda73542e5f68615a914372acb"} Nov 24 08:29:23 crc kubenswrapper[4831]: E1124 08:29:23.708486 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podUID="03d64575-b4f3-4b39-9fb8-d9597cea126c" Nov 24 08:29:23 crc kubenswrapper[4831]: I1124 08:29:23.735765 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" event={"ID":"863fe05e-d2e1-47be-b0ab-1d69c2455dc1","Type":"ContainerStarted","Data":"4530e9069c15058a30a7eae6d94b5b2d973f65ae4bb3c77e9f2d3b90851dd6cc"} Nov 24 08:29:23 crc kubenswrapper[4831]: E1124 08:29:23.737735 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podUID="82cf04aa-e08a-4626-ba4c-8c2321aaeb13" Nov 24 08:29:23 crc kubenswrapper[4831]: E1124 08:29:23.744160 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" podUID="863fe05e-d2e1-47be-b0ab-1d69c2455dc1" Nov 24 08:29:24 crc kubenswrapper[4831]: I1124 08:29:24.792826 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" event={"ID":"145ad046-da71-432f-9b5d-0f2c599c85e0","Type":"ContainerStarted","Data":"1f1474f0e039c18c11aa58ca2c456f348d8ab8390443e94e5ac8f2cac12bba00"} Nov 24 08:29:24 crc kubenswrapper[4831]: I1124 08:29:24.793343 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:24 crc kubenswrapper[4831]: E1124 08:29:24.794409 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" podUID="7531434a-d014-4bcb-9765-dd8ca784de62" Nov 24 08:29:24 crc kubenswrapper[4831]: E1124 08:29:24.795253 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podUID="03d64575-b4f3-4b39-9fb8-d9597cea126c" Nov 24 08:29:24 crc kubenswrapper[4831]: E1124 08:29:24.795270 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" podUID="863fe05e-d2e1-47be-b0ab-1d69c2455dc1" Nov 24 08:29:24 crc kubenswrapper[4831]: E1124 08:29:24.796430 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" podUID="14cc197a-4c42-46dc-89c6-941dd1997aa8" Nov 24 08:29:24 crc kubenswrapper[4831]: I1124 08:29:24.871000 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" podStartSLOduration=5.870981446 podStartE2EDuration="5.870981446s" podCreationTimestamp="2025-11-24 08:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:29:24.864343158 +0000 UTC m=+838.739488311" watchObservedRunningTime="2025-11-24 08:29:24.870981446 +0000 UTC m=+838.746126599" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.290996 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.294122 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.321248 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.388286 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp45m\" (UniqueName: \"kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.388877 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.388938 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.490000 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.490078 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp45m\" (UniqueName: \"kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.490139 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.490720 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.490982 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.514136 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp45m\" (UniqueName: \"kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m\") pod \"certified-operators-z2zfx\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:28 crc kubenswrapper[4831]: I1124 08:29:28.619114 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:31 crc kubenswrapper[4831]: I1124 08:29:31.361276 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5dbdc8d64b-pcjw6" Nov 24 08:29:36 crc kubenswrapper[4831]: E1124 08:29:36.381612 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9" Nov 24 08:29:36 crc kubenswrapper[4831]: E1124 08:29:36.382453 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:848f4c43c6bdd4e33e3ce1d147a85b9b6a6124a150bd5155dce421ef539259e9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vgtf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-598f69df5d-rshdg_openstack-operators(e94a7743-730d-4259-a584-59d1034cd654): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:37 crc kubenswrapper[4831]: E1124 08:29:37.635261 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a" Nov 24 08:29:37 crc kubenswrapper[4831]: E1124 08:29:37.635699 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:b749a5dd8bc718875c3f5e81b38d54d003be77ab92de4a3e9f9595566496a58a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nnjtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58f887965d-zftw6_openstack-operators(021fa54a-79c0-483a-96af-9f154175eb2f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:38 crc kubenswrapper[4831]: E1124 08:29:38.218296 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c" Nov 24 08:29:38 crc kubenswrapper[4831]: E1124 08:29:38.218528 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:4094e7fc11a33e8e2b6768a053cafaf5b122446d23f9113d43d520cb64e9776c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fgvlk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b797b8dff-md74d_openstack-operators(e7c48858-2602-4763-b69e-96a4f7db47f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:38 crc kubenswrapper[4831]: E1124 08:29:38.806798 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377" Nov 24 08:29:38 crc kubenswrapper[4831]: E1124 08:29:38.808564 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:b582189b55fddc180a6d468c9dba7078009a693db37b4093d4ba0c99ec675377,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ffkw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-99b499f4-rrtwx_openstack-operators(4d9702a6-a116-4598-bc95-14c7023cacb0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:39 crc kubenswrapper[4831]: E1124 08:29:39.382615 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b" Nov 24 08:29:39 crc kubenswrapper[4831]: E1124 08:29:39.382814 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5lcnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-54fc5f65b7-4c6fw_openstack-operators(907add29-1677-495e-b968-432dd37c5e4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:39 crc kubenswrapper[4831]: E1124 08:29:39.875310 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f" Nov 24 08:29:39 crc kubenswrapper[4831]: E1124 08:29:39.875803 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c6405d94e56b40ef669729216ab4b9c441f34bb280902efa2940038c076b560f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mx29s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-767ccfd65f-bdfgj_openstack-operators(cdc653a8-5c91-4da4-8489-380f2fd0b05d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:41 crc kubenswrapper[4831]: E1124 08:29:41.799740 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d" Nov 24 08:29:41 crc kubenswrapper[4831]: E1124 08:29:41.799994 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l862h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-b4c496f69-9d55q_openstack-operators(0ee539a7-3c8d-47b9-bf94-bcecc7ff9333): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:42 crc kubenswrapper[4831]: E1124 08:29:42.355495 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a" Nov 24 08:29:42 crc kubenswrapper[4831]: E1124 08:29:42.355701 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwq8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7454b96578-zz4ns_openstack-operators(8bab5961-2a79-482a-8935-b3f84b6e5fbf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:42 crc kubenswrapper[4831]: E1124 08:29:42.924582 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894" Nov 24 08:29:42 crc kubenswrapper[4831]: E1124 08:29:42.924786 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t2sqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-6dd8864d7c-m5tlh_openstack-operators(054589e5-1de8-4a64-b5ad-2d1688cf8219): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:50 crc kubenswrapper[4831]: E1124 08:29:50.583997 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f" Nov 24 08:29:50 crc kubenswrapper[4831]: E1124 08:29:50.584606 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q6mbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-8c6448b9f-k5tgn_openstack-operators(03d64575-b4f3-4b39-9fb8-d9597cea126c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:50 crc kubenswrapper[4831]: E1124 08:29:50.586479 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podUID="03d64575-b4f3-4b39-9fb8-d9597cea126c" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.039804 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.039975 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qql4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-52vxr_openstack-operators(82cf04aa-e08a-4626-ba4c-8c2321aaeb13): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.041132 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podUID="82cf04aa-e08a-4626-ba4c-8c2321aaeb13" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.192414 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" podUID="e7c48858-2602-4763-b69e-96a4f7db47f5" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.390009 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" podUID="4d9702a6-a116-4598-bc95-14c7023cacb0" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.431489 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" podUID="021fa54a-79c0-483a-96af-9f154175eb2f" Nov 24 08:29:51 crc kubenswrapper[4831]: I1124 08:29:51.525150 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.533808 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" podUID="cdc653a8-5c91-4da4-8489-380f2fd0b05d" Nov 24 08:29:51 crc kubenswrapper[4831]: W1124 08:29:51.545897 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85143835_f682_4b91_9c65_72f3b0d54bb6.slice/crio-be2c30e72991df736a8c85544e3e73a8a0d394edfcf232b23d3432d753300ca7 WatchSource:0}: Error finding container be2c30e72991df736a8c85544e3e73a8a0d394edfcf232b23d3432d753300ca7: Status 404 returned error can't find the container with id be2c30e72991df736a8c85544e3e73a8a0d394edfcf232b23d3432d753300ca7 Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.602153 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" podUID="907add29-1677-495e-b968-432dd37c5e4e" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.639302 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" podUID="e94a7743-730d-4259-a584-59d1034cd654" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.673973 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" podUID="0ee539a7-3c8d-47b9-bf94-bcecc7ff9333" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.756292 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" podUID="054589e5-1de8-4a64-b5ad-2d1688cf8219" Nov 24 08:29:51 crc kubenswrapper[4831]: E1124 08:29:51.881466 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" podUID="8bab5961-2a79-482a-8935-b3f84b6e5fbf" Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.041701 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" event={"ID":"e7c48858-2602-4763-b69e-96a4f7db47f5","Type":"ContainerStarted","Data":"85dde1852e473dd3ad1fa70f946ee9a33449e4c51ae1be1ce36bda90ee451ac6"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.043660 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.044433 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" event={"ID":"ed86dbad-6726-46f7-a11f-feb0647ab7c5","Type":"ContainerStarted","Data":"08a066ec62fa6bee8905262835d2efd68fe6a2efa8c3c7449911dd19e2d7e6c8"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.045356 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" event={"ID":"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333","Type":"ContainerStarted","Data":"71aecdd70fedafa3a17511f383f25f43b9532a473a37567c917a1e261576a403"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.047140 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" event={"ID":"054589e5-1de8-4a64-b5ad-2d1688cf8219","Type":"ContainerStarted","Data":"de93ed2d24b57bf4ddacc77c5cfafdd892e9d4d95a539b54cc4ead038a0c5331"} Nov 24 08:29:52 crc kubenswrapper[4831]: E1124 08:29:52.052525 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\"" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" podUID="054589e5-1de8-4a64-b5ad-2d1688cf8219" Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.054955 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" event={"ID":"4d9702a6-a116-4598-bc95-14c7023cacb0","Type":"ContainerStarted","Data":"3768e3f98ac7f605d11ebac4b292550db11f1a52c2dd86380a2337050c3b47d7"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.062750 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerStarted","Data":"be2c30e72991df736a8c85544e3e73a8a0d394edfcf232b23d3432d753300ca7"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.071678 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" event={"ID":"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c","Type":"ContainerStarted","Data":"f8619e4b303e97a1f78fa66ea5a475d00a0d5a098e4bdabdf718d75dade3b7b2"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.122569 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" event={"ID":"cdc653a8-5c91-4da4-8489-380f2fd0b05d","Type":"ContainerStarted","Data":"3d143b426077f2f56f44fb2ca10bdb31b5bda7457bbd672f89da640ec60407b9"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.175169 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" event={"ID":"4bc8600a-62b5-4733-90cc-b08b66b605fa","Type":"ContainerStarted","Data":"42d82f8f9b6d804d436cdf2684720a3f88030d786af9892daa9f5e4990ed049c"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.189764 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" event={"ID":"021fa54a-79c0-483a-96af-9f154175eb2f","Type":"ContainerStarted","Data":"a028058fd51d1f550466abb844b17ce1257a1e1d4b479f2ae3bcc2b5036715de"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.192366 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" event={"ID":"78bce3ee-0478-4a09-9568-9e62c29dc754","Type":"ContainerStarted","Data":"e46ff760615ed4f52c20feaecfb269f0e027eafa6811c991892536468b0f3dd5"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.229930 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" event={"ID":"37a35a63-0850-4615-afe3-905c1218bbcd","Type":"ContainerStarted","Data":"deed0686d6296c1f2da8c88a4e3acc0a77683852919724745da8a39618586de6"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.233962 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" event={"ID":"e94a7743-730d-4259-a584-59d1034cd654","Type":"ContainerStarted","Data":"f15b006ba26f877f7dac25e97472d2510060e8430a08295f9ab9ffc153687e24"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.238646 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" event={"ID":"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc","Type":"ContainerStarted","Data":"c95bd2ef544449d192dec52e9e558a941b84bf2c431ffb3d4421f9f3167b66b1"} Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.244521 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" event={"ID":"8bab5961-2a79-482a-8935-b3f84b6e5fbf","Type":"ContainerStarted","Data":"6b7a5ee86d57c152d427d54f734925c6726861e0b1eb7c9153f63f9454d88829"} Nov 24 08:29:52 crc kubenswrapper[4831]: E1124 08:29:52.245937 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" podUID="8bab5961-2a79-482a-8935-b3f84b6e5fbf" Nov 24 08:29:52 crc kubenswrapper[4831]: I1124 08:29:52.264356 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" event={"ID":"907add29-1677-495e-b968-432dd37c5e4e","Type":"ContainerStarted","Data":"55eda51f471a6521e0201042f465cbc259c0b2441a2c333b2ba9ecc484bc89e5"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.277122 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" event={"ID":"4bc8600a-62b5-4733-90cc-b08b66b605fa","Type":"ContainerStarted","Data":"915662f57ad77e6c9505563bfdf7723c6048cc29be16627a76c5004c1e8acdce"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.277542 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.291754 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" event={"ID":"ed86dbad-6726-46f7-a11f-feb0647ab7c5","Type":"ContainerStarted","Data":"47a762bba9ffac1e3c110ce4ad875c5c19e88d9428c3da960d51ad897d58a1a0"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.292446 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.324529 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" event={"ID":"78bce3ee-0478-4a09-9568-9e62c29dc754","Type":"ContainerStarted","Data":"59fd01512aee0a51c7de5bd3ae057bbb9480b93ac2a2b9ae5f5ebbd0e3185c49"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.325244 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.331661 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" event={"ID":"863fe05e-d2e1-47be-b0ab-1d69c2455dc1","Type":"ContainerStarted","Data":"b9e3f06b96dc0cada4926d5acbc60758a67b1fc67f777ffa0cc0c516441b5ecd"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.331900 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.335853 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" podStartSLOduration=11.466391247 podStartE2EDuration="35.335842265s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.494804486 +0000 UTC m=+835.369949639" lastFinishedPulling="2025-11-24 08:29:45.364255504 +0000 UTC m=+859.239400657" observedRunningTime="2025-11-24 08:29:53.333390515 +0000 UTC m=+867.208535678" watchObservedRunningTime="2025-11-24 08:29:53.335842265 +0000 UTC m=+867.210987418" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.358349 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" event={"ID":"e14d1d45-f2ee-4f9d-8260-e10bf3c8333c","Type":"ContainerStarted","Data":"daad18ee0050d2151a1a81a588b2ff1422a30bd970a6f378b0056326f9a42d27"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.358484 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" podStartSLOduration=13.191874038 podStartE2EDuration="36.358468209s" podCreationTimestamp="2025-11-24 08:29:17 +0000 UTC" firstStartedPulling="2025-11-24 08:29:20.728609603 +0000 UTC m=+834.603754756" lastFinishedPulling="2025-11-24 08:29:43.895203774 +0000 UTC m=+857.770348927" observedRunningTime="2025-11-24 08:29:53.358024697 +0000 UTC m=+867.233169850" watchObservedRunningTime="2025-11-24 08:29:53.358468209 +0000 UTC m=+867.233613352" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.358971 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.376965 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" event={"ID":"14cc197a-4c42-46dc-89c6-941dd1997aa8","Type":"ContainerStarted","Data":"9c28d56c58f8a596b0b2acd98440d5b2c9e7c15e978512c07b603842643eb24e"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.377733 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.398356 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" podStartSLOduration=5.970961905 podStartE2EDuration="35.398339326s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.833894762 +0000 UTC m=+835.709039915" lastFinishedPulling="2025-11-24 08:29:51.261272183 +0000 UTC m=+865.136417336" observedRunningTime="2025-11-24 08:29:53.393777706 +0000 UTC m=+867.268922859" watchObservedRunningTime="2025-11-24 08:29:53.398339326 +0000 UTC m=+867.273484479" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.408902 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" event={"ID":"156abc13-0fab-4e2c-a786-98ddf7c340b5","Type":"ContainerStarted","Data":"eee9ded61f48a7372fc575feabe140e0e051e9fe83b0c83e3d4c3952a6a6ea5d"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.409587 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.419689 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" event={"ID":"af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc","Type":"ContainerStarted","Data":"e156818719bd5dd8d9f77ff6892247d6d9c001ea5319c12496acc457adbe43e4"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.419830 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.428663 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" podStartSLOduration=12.559306245 podStartE2EDuration="36.428642229s" podCreationTimestamp="2025-11-24 08:29:17 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.49492054 +0000 UTC m=+835.370065693" lastFinishedPulling="2025-11-24 08:29:45.364256524 +0000 UTC m=+859.239401677" observedRunningTime="2025-11-24 08:29:53.424692406 +0000 UTC m=+867.299837589" watchObservedRunningTime="2025-11-24 08:29:53.428642229 +0000 UTC m=+867.303787382" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.456251 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" event={"ID":"37a35a63-0850-4615-afe3-905c1218bbcd","Type":"ContainerStarted","Data":"a25fd8720ee375d64aa905336630edeb6f0c405ab16b75a829e9164e491b5d12"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.456498 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.461357 4831 generic.go:334] "Generic (PLEG): container finished" podID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerID="cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad" exitCode=0 Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.461415 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerDied","Data":"cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.463689 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" event={"ID":"905703f5-8abe-45db-8e0e-a01369ef3413","Type":"ContainerStarted","Data":"4ad56bc6e020a5373cfd4fb337607aa9a6103c9bb672a4b5b20876e0920c0bf4"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.464101 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.477001 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" event={"ID":"7531434a-d014-4bcb-9765-dd8ca784de62","Type":"ContainerStarted","Data":"365a838a8fd6c4082dc1f118977b64c4dc34e0f599b8b75b4167633e303c54d9"} Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.477403 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.511267 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" podStartSLOduration=6.181485236 podStartE2EDuration="35.511250753s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.931480715 +0000 UTC m=+835.806625868" lastFinishedPulling="2025-11-24 08:29:51.261246222 +0000 UTC m=+865.136391385" observedRunningTime="2025-11-24 08:29:53.488745642 +0000 UTC m=+867.363890795" watchObservedRunningTime="2025-11-24 08:29:53.511250753 +0000 UTC m=+867.386395906" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.511853 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" podStartSLOduration=11.850534385 podStartE2EDuration="36.51184651s" podCreationTimestamp="2025-11-24 08:29:17 +0000 UTC" firstStartedPulling="2025-11-24 08:29:20.702900238 +0000 UTC m=+834.578045391" lastFinishedPulling="2025-11-24 08:29:45.364212363 +0000 UTC m=+859.239357516" observedRunningTime="2025-11-24 08:29:53.506659882 +0000 UTC m=+867.381805055" watchObservedRunningTime="2025-11-24 08:29:53.51184651 +0000 UTC m=+867.386991653" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.540585 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" podStartSLOduration=11.663707202 podStartE2EDuration="35.540567388s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.487413658 +0000 UTC m=+835.362558811" lastFinishedPulling="2025-11-24 08:29:45.364273844 +0000 UTC m=+859.239418997" observedRunningTime="2025-11-24 08:29:53.534436913 +0000 UTC m=+867.409582076" watchObservedRunningTime="2025-11-24 08:29:53.540567388 +0000 UTC m=+867.415712541" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.574192 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" podStartSLOduration=12.958745922 podStartE2EDuration="35.574168296s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.279729279 +0000 UTC m=+835.154874422" lastFinishedPulling="2025-11-24 08:29:43.895151603 +0000 UTC m=+857.770296796" observedRunningTime="2025-11-24 08:29:53.573563298 +0000 UTC m=+867.448708451" watchObservedRunningTime="2025-11-24 08:29:53.574168296 +0000 UTC m=+867.449313449" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.615008 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" podStartSLOduration=12.021951258 podStartE2EDuration="35.614992999s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.771172432 +0000 UTC m=+835.646317585" lastFinishedPulling="2025-11-24 08:29:45.364214163 +0000 UTC m=+859.239359326" observedRunningTime="2025-11-24 08:29:53.613341762 +0000 UTC m=+867.488486925" watchObservedRunningTime="2025-11-24 08:29:53.614992999 +0000 UTC m=+867.490138152" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.703929 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" podStartSLOduration=11.661141295 podStartE2EDuration="35.703905872s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.321493617 +0000 UTC m=+835.196638770" lastFinishedPulling="2025-11-24 08:29:45.364258194 +0000 UTC m=+859.239403347" observedRunningTime="2025-11-24 08:29:53.697474149 +0000 UTC m=+867.572619312" watchObservedRunningTime="2025-11-24 08:29:53.703905872 +0000 UTC m=+867.579051025" Nov 24 08:29:53 crc kubenswrapper[4831]: I1124 08:29:53.790215 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" podStartSLOduration=6.453566668 podStartE2EDuration="35.790199481s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.93167207 +0000 UTC m=+835.806817223" lastFinishedPulling="2025-11-24 08:29:51.268304873 +0000 UTC m=+865.143450036" observedRunningTime="2025-11-24 08:29:53.764245502 +0000 UTC m=+867.639390675" watchObservedRunningTime="2025-11-24 08:29:53.790199481 +0000 UTC m=+867.665344624" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.491396 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" event={"ID":"4d9702a6-a116-4598-bc95-14c7023cacb0","Type":"ContainerStarted","Data":"0d33637197c4b636596d157c00afb754892497fcf9bfabf103785a0aa7afb0bb"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.491712 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.497677 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" event={"ID":"cdc653a8-5c91-4da4-8489-380f2fd0b05d","Type":"ContainerStarted","Data":"2493ee3b847bd438a5468d2393700072ae7712e8439f38c6c252c0f7030b4554"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.497784 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.506071 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" event={"ID":"e7c48858-2602-4763-b69e-96a4f7db47f5","Type":"ContainerStarted","Data":"ee54d69215cf5bfeefc15c1c142866eb28ce34ec2277bad9d2ae3938de666db4"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.506228 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.509620 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" event={"ID":"e94a7743-730d-4259-a584-59d1034cd654","Type":"ContainerStarted","Data":"dd910eb0fac07a2f9d8598a98980b06b8d1874f4ddfd78b55e72f984d80a6bb9"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.509741 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.511040 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" event={"ID":"021fa54a-79c0-483a-96af-9f154175eb2f","Type":"ContainerStarted","Data":"1f85fcf4da74d9d2fb68fc66e25b15b29e0e74d44162a76efa6a0efcac0f1872"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.511214 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.512360 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" event={"ID":"0ee539a7-3c8d-47b9-bf94-bcecc7ff9333","Type":"ContainerStarted","Data":"e26fc7c8ece920407d76dcb121e27362fd0c04f017f21620e677ada7e3cf4cae"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.512434 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.517850 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" event={"ID":"8bab5961-2a79-482a-8935-b3f84b6e5fbf","Type":"ContainerStarted","Data":"8a847f40b8b00edd618c05065e3ee6612d6e19eac429437e4fdeff6c56d9988c"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.518078 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.523872 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" event={"ID":"905703f5-8abe-45db-8e0e-a01369ef3413","Type":"ContainerStarted","Data":"2c8ca12b58328a33d49833e59d55f698e87ddc3518b5cd9c28e07d831b94d31f"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.530920 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" event={"ID":"907add29-1677-495e-b968-432dd37c5e4e","Type":"ContainerStarted","Data":"cce587242aa504aad137f79578da5a11b8801217cd5769d26f518bda02de8a0d"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.530977 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.537234 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" event={"ID":"156abc13-0fab-4e2c-a786-98ddf7c340b5","Type":"ContainerStarted","Data":"4763b94055d550774a9eeaca4a2124dce982e1322968c00832ed9606352a4988"} Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.602773 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" podStartSLOduration=5.241170879 podStartE2EDuration="36.602756624s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.663024582 +0000 UTC m=+835.538169725" lastFinishedPulling="2025-11-24 08:29:53.024610317 +0000 UTC m=+866.899755470" observedRunningTime="2025-11-24 08:29:54.595310392 +0000 UTC m=+868.470455545" watchObservedRunningTime="2025-11-24 08:29:54.602756624 +0000 UTC m=+868.477901777" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.603158 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" podStartSLOduration=5.061827604 podStartE2EDuration="36.603154136s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.928222293 +0000 UTC m=+835.803367446" lastFinishedPulling="2025-11-24 08:29:53.469548825 +0000 UTC m=+867.344693978" observedRunningTime="2025-11-24 08:29:54.548992022 +0000 UTC m=+868.424137195" watchObservedRunningTime="2025-11-24 08:29:54.603154136 +0000 UTC m=+868.478299289" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.640622 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" podStartSLOduration=4.8242641729999995 podStartE2EDuration="36.640607453s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.479616248 +0000 UTC m=+835.354761401" lastFinishedPulling="2025-11-24 08:29:53.295959528 +0000 UTC m=+867.171104681" observedRunningTime="2025-11-24 08:29:54.638378989 +0000 UTC m=+868.513524152" watchObservedRunningTime="2025-11-24 08:29:54.640607453 +0000 UTC m=+868.515752606" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.756906 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" podStartSLOduration=3.485471085 podStartE2EDuration="36.756889666s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:20.709494034 +0000 UTC m=+834.584639187" lastFinishedPulling="2025-11-24 08:29:53.980912615 +0000 UTC m=+867.856057768" observedRunningTime="2025-11-24 08:29:54.713337605 +0000 UTC m=+868.588482768" watchObservedRunningTime="2025-11-24 08:29:54.756889666 +0000 UTC m=+868.632034809" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.760306 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" podStartSLOduration=4.643889532 podStartE2EDuration="36.760296323s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.338087865 +0000 UTC m=+835.213233008" lastFinishedPulling="2025-11-24 08:29:53.454494646 +0000 UTC m=+867.329639799" observedRunningTime="2025-11-24 08:29:54.749062753 +0000 UTC m=+868.624207916" watchObservedRunningTime="2025-11-24 08:29:54.760296323 +0000 UTC m=+868.635441476" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.793572 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" podStartSLOduration=6.122248913 podStartE2EDuration="37.793550571s" podCreationTimestamp="2025-11-24 08:29:17 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.479037171 +0000 UTC m=+835.354182334" lastFinishedPulling="2025-11-24 08:29:53.150338839 +0000 UTC m=+867.025483992" observedRunningTime="2025-11-24 08:29:54.787519439 +0000 UTC m=+868.662664592" watchObservedRunningTime="2025-11-24 08:29:54.793550571 +0000 UTC m=+868.668695724" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.832137 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" podStartSLOduration=5.046971569 podStartE2EDuration="36.83211951s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.512209397 +0000 UTC m=+835.387354550" lastFinishedPulling="2025-11-24 08:29:53.297357338 +0000 UTC m=+867.172502491" observedRunningTime="2025-11-24 08:29:54.831161073 +0000 UTC m=+868.706306236" watchObservedRunningTime="2025-11-24 08:29:54.83211951 +0000 UTC m=+868.707264663" Nov 24 08:29:54 crc kubenswrapper[4831]: I1124 08:29:54.854037 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" podStartSLOduration=5.38044037 podStartE2EDuration="36.854018684s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.8320617 +0000 UTC m=+835.707206853" lastFinishedPulling="2025-11-24 08:29:53.305640014 +0000 UTC m=+867.180785167" observedRunningTime="2025-11-24 08:29:54.849982649 +0000 UTC m=+868.725127802" watchObservedRunningTime="2025-11-24 08:29:54.854018684 +0000 UTC m=+868.729163837" Nov 24 08:29:55 crc kubenswrapper[4831]: I1124 08:29:55.551436 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" event={"ID":"054589e5-1de8-4a64-b5ad-2d1688cf8219","Type":"ContainerStarted","Data":"b4cee90a859efd33ffab9d86b9d1174f99ce482aaec1723b4e214242a29246c8"} Nov 24 08:29:55 crc kubenswrapper[4831]: I1124 08:29:55.552450 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:29:55 crc kubenswrapper[4831]: I1124 08:29:55.554376 4831 generic.go:334] "Generic (PLEG): container finished" podID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerID="e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc" exitCode=0 Nov 24 08:29:55 crc kubenswrapper[4831]: I1124 08:29:55.555534 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerDied","Data":"e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc"} Nov 24 08:29:55 crc kubenswrapper[4831]: I1124 08:29:55.595331 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" podStartSLOduration=6.982954195 podStartE2EDuration="37.595295386s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:23.367856164 +0000 UTC m=+837.243001317" lastFinishedPulling="2025-11-24 08:29:53.980197355 +0000 UTC m=+867.855342508" observedRunningTime="2025-11-24 08:29:55.576620514 +0000 UTC m=+869.451765677" watchObservedRunningTime="2025-11-24 08:29:55.595295386 +0000 UTC m=+869.470440539" Nov 24 08:29:56 crc kubenswrapper[4831]: I1124 08:29:56.563755 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerStarted","Data":"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e"} Nov 24 08:29:56 crc kubenswrapper[4831]: I1124 08:29:56.606000 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z2zfx" podStartSLOduration=26.125459132 podStartE2EDuration="28.605979183s" podCreationTimestamp="2025-11-24 08:29:28 +0000 UTC" firstStartedPulling="2025-11-24 08:29:53.470278905 +0000 UTC m=+867.345424058" lastFinishedPulling="2025-11-24 08:29:55.950798936 +0000 UTC m=+869.825944109" observedRunningTime="2025-11-24 08:29:56.582306409 +0000 UTC m=+870.457451562" watchObservedRunningTime="2025-11-24 08:29:56.605979183 +0000 UTC m=+870.481124326" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.340257 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6498cbf48f-994pv" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.505806 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7969689c84-xcdjn" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.535929 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-75fb479bcc-x7xzk" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.590118 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-56f54d6746-ppdhf" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.612525 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-598f69df5d-rshdg" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.620103 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.620177 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.676978 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-767ccfd65f-bdfgj" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.828089 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58f887965d-zftw6" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.907586 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-54b5986bb8-zp88s" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.936633 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-cfbb9c588-n45zz" Nov 24 08:29:58 crc kubenswrapper[4831]: I1124 08:29:58.983412 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78bd47f458-hv6f6" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.053999 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-99b499f4-rrtwx" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.094346 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-54cfbf4c7d-n8nkt" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.385752 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-54fc5f65b7-4c6fw" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.679651 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-z2zfx" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="registry-server" probeResult="failure" output=< Nov 24 08:29:59 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:29:59 crc kubenswrapper[4831]: > Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.705029 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6d4bf84b58-4s2q2" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.723686 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-b4c496f69-9d55q" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.728798 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-d656998f4-85c8r" Nov 24 08:29:59 crc kubenswrapper[4831]: I1124 08:29:59.765061 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b797b8dff-md74d" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.160128 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6"] Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.161384 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.163770 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.164105 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.174419 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6"] Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.194750 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7865\" (UniqueName: \"kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.194813 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.194876 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.296382 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.296448 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7865\" (UniqueName: \"kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.296494 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.297444 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.305402 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.312075 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7865\" (UniqueName: \"kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865\") pod \"collect-profiles-29399550-twhh6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.486843 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.826660 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n" Nov 24 08:30:00 crc kubenswrapper[4831]: I1124 08:30:00.945904 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6"] Nov 24 08:30:01 crc kubenswrapper[4831]: I1124 08:30:01.607091 4831 generic.go:334] "Generic (PLEG): container finished" podID="d1a4e6cd-e6b2-4024-a673-1379953efab6" containerID="4d362341de33daeb8f54789a645ecd21e7717b672276c9c20e5ae998c93344e8" exitCode=0 Nov 24 08:30:01 crc kubenswrapper[4831]: I1124 08:30:01.607215 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" event={"ID":"d1a4e6cd-e6b2-4024-a673-1379953efab6","Type":"ContainerDied","Data":"4d362341de33daeb8f54789a645ecd21e7717b672276c9c20e5ae998c93344e8"} Nov 24 08:30:01 crc kubenswrapper[4831]: I1124 08:30:01.607696 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" event={"ID":"d1a4e6cd-e6b2-4024-a673-1379953efab6","Type":"ContainerStarted","Data":"7d946540e07d4d4b9e1cb15ee7f52c7a025d7a817a47a381a0e3cb0574ca3edc"} Nov 24 08:30:01 crc kubenswrapper[4831]: E1124 08:30:01.895408 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podUID="82cf04aa-e08a-4626-ba4c-8c2321aaeb13" Nov 24 08:30:02 crc kubenswrapper[4831]: I1124 08:30:02.352969 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6dd8864d7c-m5tlh" Nov 24 08:30:02 crc kubenswrapper[4831]: I1124 08:30:02.891826 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.048183 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7865\" (UniqueName: \"kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865\") pod \"d1a4e6cd-e6b2-4024-a673-1379953efab6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.048296 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume\") pod \"d1a4e6cd-e6b2-4024-a673-1379953efab6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.048409 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume\") pod \"d1a4e6cd-e6b2-4024-a673-1379953efab6\" (UID: \"d1a4e6cd-e6b2-4024-a673-1379953efab6\") " Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.049632 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume" (OuterVolumeSpecName: "config-volume") pod "d1a4e6cd-e6b2-4024-a673-1379953efab6" (UID: "d1a4e6cd-e6b2-4024-a673-1379953efab6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.052847 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865" (OuterVolumeSpecName: "kube-api-access-x7865") pod "d1a4e6cd-e6b2-4024-a673-1379953efab6" (UID: "d1a4e6cd-e6b2-4024-a673-1379953efab6"). InnerVolumeSpecName "kube-api-access-x7865". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.053035 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d1a4e6cd-e6b2-4024-a673-1379953efab6" (UID: "d1a4e6cd-e6b2-4024-a673-1379953efab6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.150903 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7865\" (UniqueName: \"kubernetes.io/projected/d1a4e6cd-e6b2-4024-a673-1379953efab6-kube-api-access-x7865\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.150944 4831 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1a4e6cd-e6b2-4024-a673-1379953efab6-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.150954 4831 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a4e6cd-e6b2-4024-a673-1379953efab6-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.627048 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" event={"ID":"d1a4e6cd-e6b2-4024-a673-1379953efab6","Type":"ContainerDied","Data":"7d946540e07d4d4b9e1cb15ee7f52c7a025d7a817a47a381a0e3cb0574ca3edc"} Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.627093 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d946540e07d4d4b9e1cb15ee7f52c7a025d7a817a47a381a0e3cb0574ca3edc" Nov 24 08:30:03 crc kubenswrapper[4831]: I1124 08:30:03.627148 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399550-twhh6" Nov 24 08:30:05 crc kubenswrapper[4831]: E1124 08:30:05.897017 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podUID="03d64575-b4f3-4b39-9fb8-d9597cea126c" Nov 24 08:30:08 crc kubenswrapper[4831]: I1124 08:30:08.670368 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:30:08 crc kubenswrapper[4831]: I1124 08:30:08.715668 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:30:08 crc kubenswrapper[4831]: I1124 08:30:08.775492 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7454b96578-zz4ns" Nov 24 08:30:08 crc kubenswrapper[4831]: I1124 08:30:08.902180 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:30:10 crc kubenswrapper[4831]: I1124 08:30:10.683152 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z2zfx" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="registry-server" containerID="cri-o://9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e" gracePeriod=2 Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.073606 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.168964 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities\") pod \"85143835-f682-4b91-9c65-72f3b0d54bb6\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.169128 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content\") pod \"85143835-f682-4b91-9c65-72f3b0d54bb6\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.169188 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp45m\" (UniqueName: \"kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m\") pod \"85143835-f682-4b91-9c65-72f3b0d54bb6\" (UID: \"85143835-f682-4b91-9c65-72f3b0d54bb6\") " Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.170138 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities" (OuterVolumeSpecName: "utilities") pod "85143835-f682-4b91-9c65-72f3b0d54bb6" (UID: "85143835-f682-4b91-9c65-72f3b0d54bb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.175100 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m" (OuterVolumeSpecName: "kube-api-access-fp45m") pod "85143835-f682-4b91-9c65-72f3b0d54bb6" (UID: "85143835-f682-4b91-9c65-72f3b0d54bb6"). InnerVolumeSpecName "kube-api-access-fp45m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.222283 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85143835-f682-4b91-9c65-72f3b0d54bb6" (UID: "85143835-f682-4b91-9c65-72f3b0d54bb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.273132 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp45m\" (UniqueName: \"kubernetes.io/projected/85143835-f682-4b91-9c65-72f3b0d54bb6-kube-api-access-fp45m\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.273172 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.273186 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85143835-f682-4b91-9c65-72f3b0d54bb6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.695480 4831 generic.go:334] "Generic (PLEG): container finished" podID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerID="9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e" exitCode=0 Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.695546 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerDied","Data":"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e"} Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.697622 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z2zfx" event={"ID":"85143835-f682-4b91-9c65-72f3b0d54bb6","Type":"ContainerDied","Data":"be2c30e72991df736a8c85544e3e73a8a0d394edfcf232b23d3432d753300ca7"} Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.695566 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z2zfx" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.697688 4831 scope.go:117] "RemoveContainer" containerID="9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.744019 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.746840 4831 scope.go:117] "RemoveContainer" containerID="e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.747497 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z2zfx"] Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.786987 4831 scope.go:117] "RemoveContainer" containerID="cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.813249 4831 scope.go:117] "RemoveContainer" containerID="9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e" Nov 24 08:30:11 crc kubenswrapper[4831]: E1124 08:30:11.813798 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e\": container with ID starting with 9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e not found: ID does not exist" containerID="9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.813858 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e"} err="failed to get container status \"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e\": rpc error: code = NotFound desc = could not find container \"9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e\": container with ID starting with 9bcbc91557ca999b6bc6fec65e0db79ac976cf046958553894e215216b8c9f9e not found: ID does not exist" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.813896 4831 scope.go:117] "RemoveContainer" containerID="e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc" Nov 24 08:30:11 crc kubenswrapper[4831]: E1124 08:30:11.814493 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc\": container with ID starting with e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc not found: ID does not exist" containerID="e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.814528 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc"} err="failed to get container status \"e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc\": rpc error: code = NotFound desc = could not find container \"e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc\": container with ID starting with e94a39181900df872f9a0644527d0b78e9bb1807eb9dc42607a8513607fbd4dc not found: ID does not exist" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.814556 4831 scope.go:117] "RemoveContainer" containerID="cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad" Nov 24 08:30:11 crc kubenswrapper[4831]: E1124 08:30:11.814988 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad\": container with ID starting with cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad not found: ID does not exist" containerID="cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad" Nov 24 08:30:11 crc kubenswrapper[4831]: I1124 08:30:11.815018 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad"} err="failed to get container status \"cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad\": rpc error: code = NotFound desc = could not find container \"cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad\": container with ID starting with cc651fa142ea8fc753b1f7902bec14f1615ceda2c7e8325843b5c3fe1da1e5ad not found: ID does not exist" Nov 24 08:30:12 crc kubenswrapper[4831]: I1124 08:30:12.909290 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" path="/var/lib/kubelet/pods/85143835-f682-4b91-9c65-72f3b0d54bb6/volumes" Nov 24 08:30:13 crc kubenswrapper[4831]: I1124 08:30:13.717157 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" event={"ID":"82cf04aa-e08a-4626-ba4c-8c2321aaeb13","Type":"ContainerStarted","Data":"b9163b136d7fc7b74d9225bf5626a84a248430271580407cc228d74eaba9c3ce"} Nov 24 08:30:13 crc kubenswrapper[4831]: I1124 08:30:13.732119 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-52vxr" podStartSLOduration=3.219858672 podStartE2EDuration="54.732098273s" podCreationTimestamp="2025-11-24 08:29:19 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.894331157 +0000 UTC m=+835.769476310" lastFinishedPulling="2025-11-24 08:30:13.406570748 +0000 UTC m=+887.281715911" observedRunningTime="2025-11-24 08:30:13.731658011 +0000 UTC m=+887.606803214" watchObservedRunningTime="2025-11-24 08:30:13.732098273 +0000 UTC m=+887.607243426" Nov 24 08:30:19 crc kubenswrapper[4831]: I1124 08:30:19.770782 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" event={"ID":"03d64575-b4f3-4b39-9fb8-d9597cea126c","Type":"ContainerStarted","Data":"48ca4611387c4f89fb892370ee11db1ca653c0ab0ec3cbd2f68fc157352deb3a"} Nov 24 08:30:19 crc kubenswrapper[4831]: I1124 08:30:19.772492 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:30:19 crc kubenswrapper[4831]: I1124 08:30:19.806927 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" podStartSLOduration=4.373167194 podStartE2EDuration="1m1.806903437s" podCreationTimestamp="2025-11-24 08:29:18 +0000 UTC" firstStartedPulling="2025-11-24 08:29:21.95188162 +0000 UTC m=+835.827026773" lastFinishedPulling="2025-11-24 08:30:19.385617863 +0000 UTC m=+893.260763016" observedRunningTime="2025-11-24 08:30:19.794193795 +0000 UTC m=+893.669338958" watchObservedRunningTime="2025-11-24 08:30:19.806903437 +0000 UTC m=+893.682048600" Nov 24 08:30:28 crc kubenswrapper[4831]: I1124 08:30:28.401981 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:30:28 crc kubenswrapper[4831]: I1124 08:30:28.402816 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:30:29 crc kubenswrapper[4831]: I1124 08:30:29.813742 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-8c6448b9f-k5tgn" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.857647 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:30:46 crc kubenswrapper[4831]: E1124 08:30:46.858529 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="extract-utilities" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858609 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="extract-utilities" Nov 24 08:30:46 crc kubenswrapper[4831]: E1124 08:30:46.858634 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="extract-content" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858640 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="extract-content" Nov 24 08:30:46 crc kubenswrapper[4831]: E1124 08:30:46.858657 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a4e6cd-e6b2-4024-a673-1379953efab6" containerName="collect-profiles" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858663 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a4e6cd-e6b2-4024-a673-1379953efab6" containerName="collect-profiles" Nov 24 08:30:46 crc kubenswrapper[4831]: E1124 08:30:46.858689 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="registry-server" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858695 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="registry-server" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858827 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a4e6cd-e6b2-4024-a673-1379953efab6" containerName="collect-profiles" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.858846 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="85143835-f682-4b91-9c65-72f3b0d54bb6" containerName="registry-server" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.859573 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.864196 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.864435 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-kntgv" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.864624 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.864753 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 24 08:30:46 crc kubenswrapper[4831]: I1124 08:30:46.869205 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.014094 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.025275 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.029714 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.037958 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.038076 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dls7n\" (UniqueName: \"kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.047399 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.139293 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.139361 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.139406 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dls7n\" (UniqueName: \"kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.139433 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.139454 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bf9\" (UniqueName: \"kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.140298 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.179480 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dls7n\" (UniqueName: \"kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n\") pod \"dnsmasq-dns-675f4bcbfc-9kfs4\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.241531 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.242631 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.242778 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bf9\" (UniqueName: \"kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.243045 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.244036 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.265121 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bf9\" (UniqueName: \"kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9\") pod \"dnsmasq-dns-78dd6ddcc-55s66\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.338852 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.478077 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.823130 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:30:47 crc kubenswrapper[4831]: I1124 08:30:47.913307 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:30:47 crc kubenswrapper[4831]: W1124 08:30:47.921180 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod296a02fd_2985_4785_9168_b162584a61db.slice/crio-d24409a2af32c75528c1142e25f1dd93dcc6651642ac775c71219b5c35995a87 WatchSource:0}: Error finding container d24409a2af32c75528c1142e25f1dd93dcc6651642ac775c71219b5c35995a87: Status 404 returned error can't find the container with id d24409a2af32c75528c1142e25f1dd93dcc6651642ac775c71219b5c35995a87 Nov 24 08:30:48 crc kubenswrapper[4831]: I1124 08:30:48.017169 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" event={"ID":"296a02fd-2985-4785-9168-b162584a61db","Type":"ContainerStarted","Data":"d24409a2af32c75528c1142e25f1dd93dcc6651642ac775c71219b5c35995a87"} Nov 24 08:30:48 crc kubenswrapper[4831]: I1124 08:30:48.018773 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" event={"ID":"b288352a-75b0-46dd-b8ea-e48f37d0e092","Type":"ContainerStarted","Data":"8f646c9fba7aa883934709fb5ecd1673b91aa6b293a8f6cab4b2893d9e8b52be"} Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.104691 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.136520 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.137635 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.155220 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.303090 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.304664 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2829\" (UniqueName: \"kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.304713 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.406036 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.406110 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2829\" (UniqueName: \"kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.406144 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.407057 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.407089 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.430960 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2829\" (UniqueName: \"kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829\") pod \"dnsmasq-dns-666b6646f7-vqbbk\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.448652 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.462432 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.490359 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.495860 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.510881 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.608903 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ks9\" (UniqueName: \"kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.609313 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.609379 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.712851 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ks9\" (UniqueName: \"kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.712926 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.712962 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.719274 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.719482 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.766462 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ks9\" (UniqueName: \"kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9\") pod \"dnsmasq-dns-57d769cc4f-9slcr\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.840439 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:30:50 crc kubenswrapper[4831]: I1124 08:30:50.979051 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:30:50 crc kubenswrapper[4831]: W1124 08:30:50.990482 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4528c376_737c_418e_9950_5b82044254c0.slice/crio-585d0bbc56b3faf1f5e53a6297fe382fefe9780615381b5d2f8a652ea41b61ab WatchSource:0}: Error finding container 585d0bbc56b3faf1f5e53a6297fe382fefe9780615381b5d2f8a652ea41b61ab: Status 404 returned error can't find the container with id 585d0bbc56b3faf1f5e53a6297fe382fefe9780615381b5d2f8a652ea41b61ab Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.054437 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" event={"ID":"4528c376-737c-418e-9950-5b82044254c0","Type":"ContainerStarted","Data":"585d0bbc56b3faf1f5e53a6297fe382fefe9780615381b5d2f8a652ea41b61ab"} Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.313198 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.321600 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.324527 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.324945 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.326042 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7kqpn" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.326053 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.326280 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.326303 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.327482 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.327712 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.351686 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:30:51 crc kubenswrapper[4831]: W1124 08:30:51.370445 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb91ad2e2_a4e7_4cbe_a06f_cf37720b5604.slice/crio-d80b866f7c15f10f303fda9e154916f04fdcb86472d0fd078a4038708f86a079 WatchSource:0}: Error finding container d80b866f7c15f10f303fda9e154916f04fdcb86472d0fd078a4038708f86a079: Status 404 returned error can't find the container with id d80b866f7c15f10f303fda9e154916f04fdcb86472d0fd078a4038708f86a079 Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.428879 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.428936 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.428963 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.428994 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429021 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429040 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b24tl\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429108 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429155 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429177 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429196 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.429212 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.530929 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.530978 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531013 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531034 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531093 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531109 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531123 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531142 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531169 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531185 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b24tl\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.531208 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.532304 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.533665 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.534066 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.534388 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.535619 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.544628 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.545716 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.547006 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.548406 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.551070 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.555344 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b24tl\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.573690 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.638231 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.639557 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.643593 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.643669 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.643617 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.643989 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2fql7" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.644234 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.644278 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.644383 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.651961 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.669693 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744581 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4npc\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744640 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744675 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744695 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744891 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.744968 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.745021 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.745056 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.745104 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.745147 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.745239 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848004 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848053 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848088 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848132 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4npc\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848169 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848201 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848217 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848249 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848269 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848292 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.848306 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.849802 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.850074 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.850363 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.850956 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.851144 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.855232 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.868082 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.871190 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4npc\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.875682 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.879587 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.879968 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.890823 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:51 crc kubenswrapper[4831]: I1124 08:30:51.965269 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.070203 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" event={"ID":"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604","Type":"ContainerStarted","Data":"d80b866f7c15f10f303fda9e154916f04fdcb86472d0fd078a4038708f86a079"} Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.934446 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.942263 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.945748 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.950677 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.953391 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.954145 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gv97d" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.956415 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 24 08:30:52 crc kubenswrapper[4831]: I1124 08:30:52.956695 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080655 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080698 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6kbh\" (UniqueName: \"kubernetes.io/projected/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kube-api-access-x6kbh\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080772 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080801 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080817 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080841 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080954 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.080979 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185250 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185302 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185348 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185378 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185415 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185435 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185487 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.185511 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6kbh\" (UniqueName: \"kubernetes.io/projected/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kube-api-access-x6kbh\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.187002 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.192559 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.193946 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kolla-config\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.196359 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.196992 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-config-data-default\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.201300 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.233701 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6kbh\" (UniqueName: \"kubernetes.io/projected/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-kube-api-access-x6kbh\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.234713 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.291587 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2\") " pod="openstack/openstack-galera-0" Nov 24 08:30:53 crc kubenswrapper[4831]: I1124 08:30:53.579284 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.349439 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.352499 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.355829 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.355831 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.355980 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-x5btk" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.360898 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.361335 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.513739 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z42mq\" (UniqueName: \"kubernetes.io/projected/7d136719-7712-457e-a66c-d4f349c8c341-kube-api-access-z42mq\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.517490 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.517576 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.517804 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.517859 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.517890 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.518028 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.518109 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d136719-7712-457e-a66c-d4f349c8c341-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622586 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z42mq\" (UniqueName: \"kubernetes.io/projected/7d136719-7712-457e-a66c-d4f349c8c341-kube-api-access-z42mq\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622737 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622792 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622859 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622881 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.622899 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.623032 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.623072 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d136719-7712-457e-a66c-d4f349c8c341-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.623570 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7d136719-7712-457e-a66c-d4f349c8c341-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.627725 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.628357 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.628729 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7d136719-7712-457e-a66c-d4f349c8c341-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.628969 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.641267 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.648204 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z42mq\" (UniqueName: \"kubernetes.io/projected/7d136719-7712-457e-a66c-d4f349c8c341-kube-api-access-z42mq\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.656739 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d136719-7712-457e-a66c-d4f349c8c341-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.666894 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"7d136719-7712-457e-a66c-d4f349c8c341\") " pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.699572 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.701889 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.710771 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.710877 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.711213 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nqstj" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.711431 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.733861 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.825592 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8s72\" (UniqueName: \"kubernetes.io/projected/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kube-api-access-q8s72\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.826660 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-config-data\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.826758 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kolla-config\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.826786 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.826815 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.932114 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8s72\" (UniqueName: \"kubernetes.io/projected/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kube-api-access-q8s72\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.932180 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-config-data\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.932200 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kolla-config\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.932222 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.932247 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.933112 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kolla-config\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.933212 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfda28f8-5deb-43df-9548-3dabfc21d6c0-config-data\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.937799 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.955210 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfda28f8-5deb-43df-9548-3dabfc21d6c0-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:54 crc kubenswrapper[4831]: I1124 08:30:54.978853 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8s72\" (UniqueName: \"kubernetes.io/projected/cfda28f8-5deb-43df-9548-3dabfc21d6c0-kube-api-access-q8s72\") pod \"memcached-0\" (UID: \"cfda28f8-5deb-43df-9548-3dabfc21d6c0\") " pod="openstack/memcached-0" Nov 24 08:30:55 crc kubenswrapper[4831]: I1124 08:30:55.054232 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.572837 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.574272 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.576651 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7c785" Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.591265 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.762896 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkhlm\" (UniqueName: \"kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm\") pod \"kube-state-metrics-0\" (UID: \"3294a544-5049-47a8-8633-2250f41005ea\") " pod="openstack/kube-state-metrics-0" Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.864696 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkhlm\" (UniqueName: \"kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm\") pod \"kube-state-metrics-0\" (UID: \"3294a544-5049-47a8-8633-2250f41005ea\") " pod="openstack/kube-state-metrics-0" Nov 24 08:30:56 crc kubenswrapper[4831]: I1124 08:30:56.896726 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkhlm\" (UniqueName: \"kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm\") pod \"kube-state-metrics-0\" (UID: \"3294a544-5049-47a8-8633-2250f41005ea\") " pod="openstack/kube-state-metrics-0" Nov 24 08:30:57 crc kubenswrapper[4831]: I1124 08:30:57.191873 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:30:58 crc kubenswrapper[4831]: I1124 08:30:58.401086 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:30:58 crc kubenswrapper[4831]: I1124 08:30:58.401380 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.508673 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.510327 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.516639 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.517262 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.517440 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.523021 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8h2n9" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.523255 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.527739 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627159 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627272 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627502 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgcbp\" (UniqueName: \"kubernetes.io/projected/b44afa71-a830-4284-87a1-d42b96db03a9-kube-api-access-dgcbp\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627523 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627547 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627566 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-config\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627587 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.627615 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729005 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgcbp\" (UniqueName: \"kubernetes.io/projected/b44afa71-a830-4284-87a1-d42b96db03a9-kube-api-access-dgcbp\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729060 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729096 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729129 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-config\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729151 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729182 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729235 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.729276 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.738989 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.740834 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.741950 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b44afa71-a830-4284-87a1-d42b96db03a9-config\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.742542 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.749704 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.750367 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.757293 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b44afa71-a830-4284-87a1-d42b96db03a9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.772248 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgcbp\" (UniqueName: \"kubernetes.io/projected/b44afa71-a830-4284-87a1-d42b96db03a9-kube-api-access-dgcbp\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.811108 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"b44afa71-a830-4284-87a1-d42b96db03a9\") " pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:00 crc kubenswrapper[4831]: I1124 08:31:00.852972 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.870996 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p87rh"] Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.873593 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.877424 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-x6v4h" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.877634 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.877960 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.892875 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh"] Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.924165 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-4wbwz"] Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.925684 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:01 crc kubenswrapper[4831]: I1124 08:31:01.942471 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4wbwz"] Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.055649 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-log-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.055699 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-run\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.055727 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-etc-ovs\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056061 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-combined-ca-bundle\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056119 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-lib\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056253 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf8tx\" (UniqueName: \"kubernetes.io/projected/71dd0aae-5932-4334-9639-49a91209160c-kube-api-access-tf8tx\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056299 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056367 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056468 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-ovn-controller-tls-certs\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056698 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c454278-b245-4dba-a3e5-0fbb86db5ab0-scripts\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056741 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t925\" (UniqueName: \"kubernetes.io/projected/2c454278-b245-4dba-a3e5-0fbb86db5ab0-kube-api-access-4t925\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.056942 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dd0aae-5932-4334-9639-49a91209160c-scripts\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.057007 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-log\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.158195 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dd0aae-5932-4334-9639-49a91209160c-scripts\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.158251 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-log\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.158272 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-log-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.158925 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-log-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.158999 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-run\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159024 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-etc-ovs\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159047 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-run\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159085 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-combined-ca-bundle\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159238 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-lib\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf8tx\" (UniqueName: \"kubernetes.io/projected/71dd0aae-5932-4334-9639-49a91209160c-kube-api-access-tf8tx\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159312 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159347 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159382 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-ovn-controller-tls-certs\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159402 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c454278-b245-4dba-a3e5-0fbb86db5ab0-scripts\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159418 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t925\" (UniqueName: \"kubernetes.io/projected/2c454278-b245-4dba-a3e5-0fbb86db5ab0-kube-api-access-4t925\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159434 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159382 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-lib\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159617 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-etc-ovs\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.159688 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/71dd0aae-5932-4334-9639-49a91209160c-var-run-ovn\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.160101 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c454278-b245-4dba-a3e5-0fbb86db5ab0-var-log\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.162597 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71dd0aae-5932-4334-9639-49a91209160c-scripts\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.162096 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c454278-b245-4dba-a3e5-0fbb86db5ab0-scripts\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.166364 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-combined-ca-bundle\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.167060 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/71dd0aae-5932-4334-9639-49a91209160c-ovn-controller-tls-certs\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.179439 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf8tx\" (UniqueName: \"kubernetes.io/projected/71dd0aae-5932-4334-9639-49a91209160c-kube-api-access-tf8tx\") pod \"ovn-controller-p87rh\" (UID: \"71dd0aae-5932-4334-9639-49a91209160c\") " pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.186703 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t925\" (UniqueName: \"kubernetes.io/projected/2c454278-b245-4dba-a3e5-0fbb86db5ab0-kube-api-access-4t925\") pod \"ovn-controller-ovs-4wbwz\" (UID: \"2c454278-b245-4dba-a3e5-0fbb86db5ab0\") " pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.200704 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh" Nov 24 08:31:02 crc kubenswrapper[4831]: I1124 08:31:02.284142 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.697119 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.702910 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.707948 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.708206 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.708284 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.708379 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6kfvv" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.708305 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.793841 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.793883 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.793919 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.793949 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.794020 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.794044 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbmcf\" (UniqueName: \"kubernetes.io/projected/9974ab31-9887-4acf-8491-0e8b3aec7026-kube-api-access-tbmcf\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.794168 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-config\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.794225 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896212 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896268 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896343 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896374 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbmcf\" (UniqueName: \"kubernetes.io/projected/9974ab31-9887-4acf-8491-0e8b3aec7026-kube-api-access-tbmcf\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896427 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-config\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896462 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896534 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.896561 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.897308 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.897983 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-config\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.898344 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9974ab31-9887-4acf-8491-0e8b3aec7026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.899599 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.902582 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.904008 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.916567 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9974ab31-9887-4acf-8491-0e8b3aec7026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.944875 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:03 crc kubenswrapper[4831]: I1124 08:31:03.945558 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbmcf\" (UniqueName: \"kubernetes.io/projected/9974ab31-9887-4acf-8491-0e8b3aec7026-kube-api-access-tbmcf\") pod \"ovsdbserver-nb-0\" (UID: \"9974ab31-9887-4acf-8491-0e8b3aec7026\") " pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:04 crc kubenswrapper[4831]: I1124 08:31:04.032007 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:05 crc kubenswrapper[4831]: I1124 08:31:05.986165 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.377090 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.377290 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dls7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9kfs4_openstack(296a02fd-2985-4785-9168-b162584a61db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.378686 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" podUID="296a02fd-2985-4785-9168-b162584a61db" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.390226 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.390532 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h2bf9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-55s66_openstack(b288352a-75b0-46dd-b8ea-e48f37d0e092): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.391702 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" podUID="b288352a-75b0-46dd-b8ea-e48f37d0e092" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.441256 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.441456 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b2829,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-vqbbk_openstack(4528c376-737c-418e-9950-5b82044254c0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:31:06 crc kubenswrapper[4831]: E1124 08:31:06.442903 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" podUID="4528c376-737c-418e-9950-5b82044254c0" Nov 24 08:31:06 crc kubenswrapper[4831]: I1124 08:31:06.825897 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.051490 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.164845 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.209911 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cfda28f8-5deb-43df-9548-3dabfc21d6c0","Type":"ContainerStarted","Data":"bf70157cad87d08857bb3f57d6a55d47ec3230db2faa858b3a77f6671f69cbb6"} Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.211730 4831 generic.go:334] "Generic (PLEG): container finished" podID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerID="aad9024288c938941c0bef90bfd7d139de74a98c106f686d56127d941841b3d5" exitCode=0 Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.211780 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" event={"ID":"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604","Type":"ContainerDied","Data":"aad9024288c938941c0bef90bfd7d139de74a98c106f686d56127d941841b3d5"} Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.214386 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3294a544-5049-47a8-8633-2250f41005ea","Type":"ContainerStarted","Data":"9f93f8fb4442131a65010e9afa13bc22e26786a6ab33643220df40f176757af2"} Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.215555 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2","Type":"ContainerStarted","Data":"2ae94e5dcc903e80c43d43ea40ed0ef149e0df880fb870d4ef6b30c7002539f2"} Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.218480 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7d136719-7712-457e-a66c-d4f349c8c341","Type":"ContainerStarted","Data":"4ca0e4c3b3be15aa59fa992dd534f9f18653fc9fa0e57a05877c442308108599"} Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.420394 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.436460 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:31:07 crc kubenswrapper[4831]: W1124 08:31:07.444264 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8b0fe50_d340_4379_8be9_4367a9c81d9f.slice/crio-d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e WatchSource:0}: Error finding container d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e: Status 404 returned error can't find the container with id d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.445011 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh"] Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.564177 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-4wbwz"] Nov 24 08:31:07 crc kubenswrapper[4831]: W1124 08:31:07.620901 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c454278_b245_4dba_a3e5_0fbb86db5ab0.slice/crio-6b5fd53b233b4bbc0ae0267fbe43798f52db5b1682a53c3359f718087ca1b59a WatchSource:0}: Error finding container 6b5fd53b233b4bbc0ae0267fbe43798f52db5b1682a53c3359f718087ca1b59a: Status 404 returned error can't find the container with id 6b5fd53b233b4bbc0ae0267fbe43798f52db5b1682a53c3359f718087ca1b59a Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.739539 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.828238 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.902965 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc\") pod \"b288352a-75b0-46dd-b8ea-e48f37d0e092\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.903242 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config\") pod \"b288352a-75b0-46dd-b8ea-e48f37d0e092\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.903380 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2bf9\" (UniqueName: \"kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9\") pod \"b288352a-75b0-46dd-b8ea-e48f37d0e092\" (UID: \"b288352a-75b0-46dd-b8ea-e48f37d0e092\") " Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.903890 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config" (OuterVolumeSpecName: "config") pod "b288352a-75b0-46dd-b8ea-e48f37d0e092" (UID: "b288352a-75b0-46dd-b8ea-e48f37d0e092"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.904174 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.904256 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b288352a-75b0-46dd-b8ea-e48f37d0e092" (UID: "b288352a-75b0-46dd-b8ea-e48f37d0e092"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:07 crc kubenswrapper[4831]: I1124 08:31:07.906907 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9" (OuterVolumeSpecName: "kube-api-access-h2bf9") pod "b288352a-75b0-46dd-b8ea-e48f37d0e092" (UID: "b288352a-75b0-46dd-b8ea-e48f37d0e092"). InnerVolumeSpecName "kube-api-access-h2bf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.006458 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dls7n\" (UniqueName: \"kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n\") pod \"296a02fd-2985-4785-9168-b162584a61db\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.006668 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config\") pod \"296a02fd-2985-4785-9168-b162584a61db\" (UID: \"296a02fd-2985-4785-9168-b162584a61db\") " Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.007063 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2bf9\" (UniqueName: \"kubernetes.io/projected/b288352a-75b0-46dd-b8ea-e48f37d0e092-kube-api-access-h2bf9\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.007077 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b288352a-75b0-46dd-b8ea-e48f37d0e092-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.010841 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config" (OuterVolumeSpecName: "config") pod "296a02fd-2985-4785-9168-b162584a61db" (UID: "296a02fd-2985-4785-9168-b162584a61db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.012845 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n" (OuterVolumeSpecName: "kube-api-access-dls7n") pod "296a02fd-2985-4785-9168-b162584a61db" (UID: "296a02fd-2985-4785-9168-b162584a61db"). InnerVolumeSpecName "kube-api-access-dls7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.017836 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-g88jf"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.038769 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.041870 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.049521 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-g88jf"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.109340 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/296a02fd-2985-4785-9168-b162584a61db-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.109387 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dls7n\" (UniqueName: \"kubernetes.io/projected/296a02fd-2985-4785-9168-b162584a61db-kube-api-access-dls7n\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.211903 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-combined-ca-bundle\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.212539 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovn-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.212598 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovs-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.212679 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.212711 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2gc\" (UniqueName: \"kubernetes.io/projected/23b22b28-ece1-400e-9e58-a34fa7392bfe-kube-api-access-zf2gc\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.212788 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b22b28-ece1-400e-9e58-a34fa7392bfe-config\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.233379 4831 generic.go:334] "Generic (PLEG): container finished" podID="4528c376-737c-418e-9950-5b82044254c0" containerID="dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188" exitCode=0 Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.234202 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" event={"ID":"4528c376-737c-418e-9950-5b82044254c0","Type":"ContainerDied","Data":"dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.243896 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" event={"ID":"296a02fd-2985-4785-9168-b162584a61db","Type":"ContainerDied","Data":"d24409a2af32c75528c1142e25f1dd93dcc6651642ac775c71219b5c35995a87"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.244023 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9kfs4" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.247948 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" event={"ID":"b288352a-75b0-46dd-b8ea-e48f37d0e092","Type":"ContainerDied","Data":"8f646c9fba7aa883934709fb5ecd1673b91aa6b293a8f6cab4b2893d9e8b52be"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.248105 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-55s66" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.259555 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" event={"ID":"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604","Type":"ContainerStarted","Data":"f74ef8d260f27be9f62d68e1bcf97dc381e49fc59caf9472619ef4334ef75252"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.267579 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4wbwz" event={"ID":"2c454278-b245-4dba-a3e5-0fbb86db5ab0","Type":"ContainerStarted","Data":"6b5fd53b233b4bbc0ae0267fbe43798f52db5b1682a53c3359f718087ca1b59a"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.273646 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh" event={"ID":"71dd0aae-5932-4334-9639-49a91209160c","Type":"ContainerStarted","Data":"0eb885eaf64b968c6544d279e4da76234efdfdb66a01af8ea3a334f351c0671e"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.276001 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerStarted","Data":"1ce3ec7a44ca880aedd4b02baa267184e1367846c10e857c03611c30900ed1c1"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.282260 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerStarted","Data":"d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e"} Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.300182 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" podStartSLOduration=3.130641499 podStartE2EDuration="18.299524219s" podCreationTimestamp="2025-11-24 08:30:50 +0000 UTC" firstStartedPulling="2025-11-24 08:30:51.373365348 +0000 UTC m=+925.248510501" lastFinishedPulling="2025-11-24 08:31:06.542248068 +0000 UTC m=+940.417393221" observedRunningTime="2025-11-24 08:31:08.293023054 +0000 UTC m=+942.168168227" watchObservedRunningTime="2025-11-24 08:31:08.299524219 +0000 UTC m=+942.174669372" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.314523 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovs-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.314627 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.314664 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2gc\" (UniqueName: \"kubernetes.io/projected/23b22b28-ece1-400e-9e58-a34fa7392bfe-kube-api-access-zf2gc\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.314707 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b22b28-ece1-400e-9e58-a34fa7392bfe-config\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.314775 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-combined-ca-bundle\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.315631 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovn-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.316065 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovn-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.316133 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/23b22b28-ece1-400e-9e58-a34fa7392bfe-ovs-rundir\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.317864 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23b22b28-ece1-400e-9e58-a34fa7392bfe-config\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.326341 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.332673 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23b22b28-ece1-400e-9e58-a34fa7392bfe-combined-ca-bundle\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.338497 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.351853 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2gc\" (UniqueName: \"kubernetes.io/projected/23b22b28-ece1-400e-9e58-a34fa7392bfe-kube-api-access-zf2gc\") pod \"ovn-controller-metrics-g88jf\" (UID: \"23b22b28-ece1-400e-9e58-a34fa7392bfe\") " pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.355632 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.365897 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9kfs4"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.383225 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-g88jf" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.383919 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.386461 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-55s66"] Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.832825 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 24 08:31:08 crc kubenswrapper[4831]: W1124 08:31:08.876829 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9974ab31_9887_4acf_8491_0e8b3aec7026.slice/crio-5a24754bd7ffbca016349d061a7c4eb6c0d2f4923d4b6dc1d249d977e737fb8e WatchSource:0}: Error finding container 5a24754bd7ffbca016349d061a7c4eb6c0d2f4923d4b6dc1d249d977e737fb8e: Status 404 returned error can't find the container with id 5a24754bd7ffbca016349d061a7c4eb6c0d2f4923d4b6dc1d249d977e737fb8e Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.912433 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296a02fd-2985-4785-9168-b162584a61db" path="/var/lib/kubelet/pods/296a02fd-2985-4785-9168-b162584a61db/volumes" Nov 24 08:31:08 crc kubenswrapper[4831]: I1124 08:31:08.912914 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b288352a-75b0-46dd-b8ea-e48f37d0e092" path="/var/lib/kubelet/pods/b288352a-75b0-46dd-b8ea-e48f37d0e092/volumes" Nov 24 08:31:09 crc kubenswrapper[4831]: I1124 08:31:09.292277 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9974ab31-9887-4acf-8491-0e8b3aec7026","Type":"ContainerStarted","Data":"5a24754bd7ffbca016349d061a7c4eb6c0d2f4923d4b6dc1d249d977e737fb8e"} Nov 24 08:31:09 crc kubenswrapper[4831]: I1124 08:31:09.295052 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b44afa71-a830-4284-87a1-d42b96db03a9","Type":"ContainerStarted","Data":"7d84987d44b2cca676027a8fb18f2fede53bc216c631154462d25d58df738faf"} Nov 24 08:31:09 crc kubenswrapper[4831]: I1124 08:31:09.295262 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:31:15 crc kubenswrapper[4831]: I1124 08:31:15.842545 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:31:15 crc kubenswrapper[4831]: I1124 08:31:15.896450 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:31:20 crc kubenswrapper[4831]: I1124 08:31:20.054027 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-g88jf"] Nov 24 08:31:20 crc kubenswrapper[4831]: W1124 08:31:20.096397 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23b22b28_ece1_400e_9e58_a34fa7392bfe.slice/crio-679ac2d1ac05c9b763a5c7e9a2a4f3979224641de3169e80f743ae5e16123977 WatchSource:0}: Error finding container 679ac2d1ac05c9b763a5c7e9a2a4f3979224641de3169e80f743ae5e16123977: Status 404 returned error can't find the container with id 679ac2d1ac05c9b763a5c7e9a2a4f3979224641de3169e80f743ae5e16123977 Nov 24 08:31:20 crc kubenswrapper[4831]: I1124 08:31:20.399256 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-g88jf" event={"ID":"23b22b28-ece1-400e-9e58-a34fa7392bfe","Type":"ContainerStarted","Data":"679ac2d1ac05c9b763a5c7e9a2a4f3979224641de3169e80f743ae5e16123977"} Nov 24 08:31:20 crc kubenswrapper[4831]: E1124 08:31:20.456813 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 24 08:31:20 crc kubenswrapper[4831]: E1124 08:31:20.456854 4831 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Nov 24 08:31:20 crc kubenswrapper[4831]: E1124 08:31:20.456965 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wkhlm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(3294a544-5049-47a8-8633-2250f41005ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 24 08:31:20 crc kubenswrapper[4831]: E1124 08:31:20.458407 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="3294a544-5049-47a8-8633-2250f41005ea" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.417926 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7d136719-7712-457e-a66c-d4f349c8c341","Type":"ContainerStarted","Data":"3e0daa105804831c57a5b962c43faafc6f9251801146d1e368bb18cd6605ebc4"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.431111 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b44afa71-a830-4284-87a1-d42b96db03a9","Type":"ContainerStarted","Data":"36f5833c507c46d87e0717cb5b44e80975fbcb93c3a3366b5bf936b0478921fb"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.438404 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cfda28f8-5deb-43df-9548-3dabfc21d6c0","Type":"ContainerStarted","Data":"74a17c0e1f0f1ed94f99208e3d4150a10ccd6f3dfe5ea82fc52d7b8d5ecdfce1"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.438914 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.445588 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4wbwz" event={"ID":"2c454278-b245-4dba-a3e5-0fbb86db5ab0","Type":"ContainerStarted","Data":"b3b663bfe0c1fda48351a72a68b52745d2b507b72933d64914c758f65079789c"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.449900 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2","Type":"ContainerStarted","Data":"36e9e56323eb9591256bf2aef1ca328b0cbb6dc4e0f89f5493f019a97f6c1c6e"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.452330 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh" event={"ID":"71dd0aae-5932-4334-9639-49a91209160c","Type":"ContainerStarted","Data":"45ea8e52610043f54a32695263fb9fe65fc37d54158070770232dfeffab76c25"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.452749 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-p87rh" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.457549 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" event={"ID":"4528c376-737c-418e-9950-5b82044254c0","Type":"ContainerStarted","Data":"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30"} Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.457662 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="dnsmasq-dns" containerID="cri-o://a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30" gracePeriod=10 Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.457918 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.471160 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9974ab31-9887-4acf-8491-0e8b3aec7026","Type":"ContainerStarted","Data":"ea20a51cd87cd399bbb0433cfec2b5520e04bc5c33264c4a8f5cca43b176975e"} Nov 24 08:31:21 crc kubenswrapper[4831]: E1124 08:31:21.473020 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="3294a544-5049-47a8-8633-2250f41005ea" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.485182 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.913551316 podStartE2EDuration="27.485159369s" podCreationTimestamp="2025-11-24 08:30:54 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.086955909 +0000 UTC m=+940.962101062" lastFinishedPulling="2025-11-24 08:31:19.658563962 +0000 UTC m=+953.533709115" observedRunningTime="2025-11-24 08:31:21.484021847 +0000 UTC m=+955.359167010" watchObservedRunningTime="2025-11-24 08:31:21.485159369 +0000 UTC m=+955.360304522" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.589640 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" podStartSLOduration=-9223372005.265165 podStartE2EDuration="31.589610545s" podCreationTimestamp="2025-11-24 08:30:50 +0000 UTC" firstStartedPulling="2025-11-24 08:30:50.992713522 +0000 UTC m=+924.867858665" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:31:21.542605696 +0000 UTC m=+955.417750849" watchObservedRunningTime="2025-11-24 08:31:21.589610545 +0000 UTC m=+955.464755698" Nov 24 08:31:21 crc kubenswrapper[4831]: I1124 08:31:21.634514 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-p87rh" podStartSLOduration=7.657885401 podStartE2EDuration="20.634455553s" podCreationTimestamp="2025-11-24 08:31:01 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.465414193 +0000 UTC m=+941.340559336" lastFinishedPulling="2025-11-24 08:31:20.441984335 +0000 UTC m=+954.317129488" observedRunningTime="2025-11-24 08:31:21.627397362 +0000 UTC m=+955.502542525" watchObservedRunningTime="2025-11-24 08:31:21.634455553 +0000 UTC m=+955.509600726" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.256011 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.418616 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc\") pod \"4528c376-737c-418e-9950-5b82044254c0\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.418663 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2829\" (UniqueName: \"kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829\") pod \"4528c376-737c-418e-9950-5b82044254c0\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.418829 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config\") pod \"4528c376-737c-418e-9950-5b82044254c0\" (UID: \"4528c376-737c-418e-9950-5b82044254c0\") " Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.429038 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829" (OuterVolumeSpecName: "kube-api-access-b2829") pod "4528c376-737c-418e-9950-5b82044254c0" (UID: "4528c376-737c-418e-9950-5b82044254c0"). InnerVolumeSpecName "kube-api-access-b2829". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.494859 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config" (OuterVolumeSpecName: "config") pod "4528c376-737c-418e-9950-5b82044254c0" (UID: "4528c376-737c-418e-9950-5b82044254c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.505026 4831 generic.go:334] "Generic (PLEG): container finished" podID="2c454278-b245-4dba-a3e5-0fbb86db5ab0" containerID="b3b663bfe0c1fda48351a72a68b52745d2b507b72933d64914c758f65079789c" exitCode=0 Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.505696 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4wbwz" event={"ID":"2c454278-b245-4dba-a3e5-0fbb86db5ab0","Type":"ContainerDied","Data":"b3b663bfe0c1fda48351a72a68b52745d2b507b72933d64914c758f65079789c"} Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.517581 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerStarted","Data":"351e96e33d34f1dd7a1d56c166ca6a8cefdb3058b5fba117be0e7a2f7ffb73d5"} Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.521069 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.521102 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2829\" (UniqueName: \"kubernetes.io/projected/4528c376-737c-418e-9950-5b82044254c0-kube-api-access-b2829\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.522492 4831 generic.go:334] "Generic (PLEG): container finished" podID="4528c376-737c-418e-9950-5b82044254c0" containerID="a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30" exitCode=0 Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.522576 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" event={"ID":"4528c376-737c-418e-9950-5b82044254c0","Type":"ContainerDied","Data":"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30"} Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.522603 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" event={"ID":"4528c376-737c-418e-9950-5b82044254c0","Type":"ContainerDied","Data":"585d0bbc56b3faf1f5e53a6297fe382fefe9780615381b5d2f8a652ea41b61ab"} Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.522620 4831 scope.go:117] "RemoveContainer" containerID="a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.522762 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-vqbbk" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.545261 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerStarted","Data":"6a1f253dccd2eceb4b96368b8a5e1b2330dfc1c9a7d3318435c861f05c2d1f8c"} Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.553258 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4528c376-737c-418e-9950-5b82044254c0" (UID: "4528c376-737c-418e-9950-5b82044254c0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.626023 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4528c376-737c-418e-9950-5b82044254c0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.636048 4831 scope.go:117] "RemoveContainer" containerID="dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.691015 4831 scope.go:117] "RemoveContainer" containerID="a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30" Nov 24 08:31:22 crc kubenswrapper[4831]: E1124 08:31:22.696098 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30\": container with ID starting with a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30 not found: ID does not exist" containerID="a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.696142 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30"} err="failed to get container status \"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30\": rpc error: code = NotFound desc = could not find container \"a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30\": container with ID starting with a97deb23d6a71212dea2b581cef27912f7a8145528a69a65d467f194bcca9f30 not found: ID does not exist" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.696171 4831 scope.go:117] "RemoveContainer" containerID="dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188" Nov 24 08:31:22 crc kubenswrapper[4831]: E1124 08:31:22.697181 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188\": container with ID starting with dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188 not found: ID does not exist" containerID="dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.697230 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188"} err="failed to get container status \"dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188\": rpc error: code = NotFound desc = could not find container \"dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188\": container with ID starting with dc0baeedd2c223a0382404a307f34591a2931989115ed7fd712a0ee535cea188 not found: ID does not exist" Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.856821 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.864051 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-vqbbk"] Nov 24 08:31:22 crc kubenswrapper[4831]: I1124 08:31:22.905560 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4528c376-737c-418e-9950-5b82044254c0" path="/var/lib/kubelet/pods/4528c376-737c-418e-9950-5b82044254c0/volumes" Nov 24 08:31:23 crc kubenswrapper[4831]: I1124 08:31:23.547171 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4wbwz" event={"ID":"2c454278-b245-4dba-a3e5-0fbb86db5ab0","Type":"ContainerStarted","Data":"fb515bf94864a30921e4f47d886fb26ff58b1027327b452945b00d88f5736e6e"} Nov 24 08:31:23 crc kubenswrapper[4831]: I1124 08:31:23.547550 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-4wbwz" event={"ID":"2c454278-b245-4dba-a3e5-0fbb86db5ab0","Type":"ContainerStarted","Data":"c23d1e92f6a62faed5af975671f46f36ef62ee6f2495149b926c1f00c5d2d3d5"} Nov 24 08:31:23 crc kubenswrapper[4831]: I1124 08:31:23.570198 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-4wbwz" podStartSLOduration=10.557373918 podStartE2EDuration="22.570176279s" podCreationTimestamp="2025-11-24 08:31:01 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.646453001 +0000 UTC m=+941.521598154" lastFinishedPulling="2025-11-24 08:31:19.659255362 +0000 UTC m=+953.534400515" observedRunningTime="2025-11-24 08:31:23.569917882 +0000 UTC m=+957.445063055" watchObservedRunningTime="2025-11-24 08:31:23.570176279 +0000 UTC m=+957.445321442" Nov 24 08:31:24 crc kubenswrapper[4831]: I1124 08:31:24.555128 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:24 crc kubenswrapper[4831]: I1124 08:31:24.555491 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:25 crc kubenswrapper[4831]: I1124 08:31:25.057816 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 24 08:31:25 crc kubenswrapper[4831]: I1124 08:31:25.564637 4831 generic.go:334] "Generic (PLEG): container finished" podID="e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2" containerID="36e9e56323eb9591256bf2aef1ca328b0cbb6dc4e0f89f5493f019a97f6c1c6e" exitCode=0 Nov 24 08:31:25 crc kubenswrapper[4831]: I1124 08:31:25.564726 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2","Type":"ContainerDied","Data":"36e9e56323eb9591256bf2aef1ca328b0cbb6dc4e0f89f5493f019a97f6c1c6e"} Nov 24 08:31:25 crc kubenswrapper[4831]: I1124 08:31:25.573968 4831 generic.go:334] "Generic (PLEG): container finished" podID="7d136719-7712-457e-a66c-d4f349c8c341" containerID="3e0daa105804831c57a5b962c43faafc6f9251801146d1e368bb18cd6605ebc4" exitCode=0 Nov 24 08:31:25 crc kubenswrapper[4831]: I1124 08:31:25.574091 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7d136719-7712-457e-a66c-d4f349c8c341","Type":"ContainerDied","Data":"3e0daa105804831c57a5b962c43faafc6f9251801146d1e368bb18cd6605ebc4"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.582465 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9974ab31-9887-4acf-8491-0e8b3aec7026","Type":"ContainerStarted","Data":"3b118b88efc3cadf162ca30e911d8b56f2d23de57c3f1782dd03d7d01593be27"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.586419 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"7d136719-7712-457e-a66c-d4f349c8c341","Type":"ContainerStarted","Data":"8e54889c362d4714b25640f4c63fab3f70a32a086ad44cbfbf4d172f33dcc1dc"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.588058 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"b44afa71-a830-4284-87a1-d42b96db03a9","Type":"ContainerStarted","Data":"f2af5a0308c27569c178bf8c75b3f586db324e80ffdd95a80e873a9feb9b5a81"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.590509 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2","Type":"ContainerStarted","Data":"32fe94b1ae00f43bf8ce0d5fe99e6de75e079803163727a9cc1cde6cc8037e8c"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.591988 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-g88jf" event={"ID":"23b22b28-ece1-400e-9e58-a34fa7392bfe","Type":"ContainerStarted","Data":"60cba76d0b09eccfd46c327915011257388ba309792a51ca8ec53aede42eb6b3"} Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.609795 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.283194428 podStartE2EDuration="24.609779669s" podCreationTimestamp="2025-11-24 08:31:02 +0000 UTC" firstStartedPulling="2025-11-24 08:31:08.89090356 +0000 UTC m=+942.766048713" lastFinishedPulling="2025-11-24 08:31:26.217488801 +0000 UTC m=+960.092633954" observedRunningTime="2025-11-24 08:31:26.608871603 +0000 UTC m=+960.484016756" watchObservedRunningTime="2025-11-24 08:31:26.609779669 +0000 UTC m=+960.484924822" Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.633510 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.4020514 podStartE2EDuration="33.633493044s" podCreationTimestamp="2025-11-24 08:30:53 +0000 UTC" firstStartedPulling="2025-11-24 08:31:06.840124936 +0000 UTC m=+940.715270089" lastFinishedPulling="2025-11-24 08:31:20.07156659 +0000 UTC m=+953.946711733" observedRunningTime="2025-11-24 08:31:26.626889586 +0000 UTC m=+960.502034749" watchObservedRunningTime="2025-11-24 08:31:26.633493044 +0000 UTC m=+960.508638197" Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.687294 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-g88jf" podStartSLOduration=13.527909863 podStartE2EDuration="19.687272977s" podCreationTimestamp="2025-11-24 08:31:07 +0000 UTC" firstStartedPulling="2025-11-24 08:31:20.098098956 +0000 UTC m=+953.973244109" lastFinishedPulling="2025-11-24 08:31:26.25746207 +0000 UTC m=+960.132607223" observedRunningTime="2025-11-24 08:31:26.653931807 +0000 UTC m=+960.529076980" watchObservedRunningTime="2025-11-24 08:31:26.687272977 +0000 UTC m=+960.562418130" Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.689276 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=10.281366047 podStartE2EDuration="27.689265924s" podCreationTimestamp="2025-11-24 08:30:59 +0000 UTC" firstStartedPulling="2025-11-24 08:31:08.8772071 +0000 UTC m=+942.752352253" lastFinishedPulling="2025-11-24 08:31:26.285106977 +0000 UTC m=+960.160252130" observedRunningTime="2025-11-24 08:31:26.684379574 +0000 UTC m=+960.559524747" watchObservedRunningTime="2025-11-24 08:31:26.689265924 +0000 UTC m=+960.564411077" Nov 24 08:31:26 crc kubenswrapper[4831]: I1124 08:31:26.717267 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.68682522 podStartE2EDuration="35.717247831s" podCreationTimestamp="2025-11-24 08:30:51 +0000 UTC" firstStartedPulling="2025-11-24 08:31:06.412431959 +0000 UTC m=+940.287577112" lastFinishedPulling="2025-11-24 08:31:20.44285455 +0000 UTC m=+954.317999723" observedRunningTime="2025-11-24 08:31:26.711467236 +0000 UTC m=+960.586612399" watchObservedRunningTime="2025-11-24 08:31:26.717247831 +0000 UTC m=+960.592392994" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.001258 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-jl98t"] Nov 24 08:31:27 crc kubenswrapper[4831]: E1124 08:31:27.001865 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="dnsmasq-dns" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.001878 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="dnsmasq-dns" Nov 24 08:31:27 crc kubenswrapper[4831]: E1124 08:31:27.001896 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="init" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.001902 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="init" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.002202 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="4528c376-737c-418e-9950-5b82044254c0" containerName="dnsmasq-dns" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.004416 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.009249 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.013986 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-jl98t"] Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.119780 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.120252 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.120360 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.120409 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78j8\" (UniqueName: \"kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.152231 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-jl98t"] Nov 24 08:31:27 crc kubenswrapper[4831]: E1124 08:31:27.153157 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-f78j8 ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" podUID="c695bfa7-8a98-45d4-97df-56955479f855" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.199058 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.200352 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.201990 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.221643 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222468 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222666 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222702 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222740 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222771 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78j8\" (UniqueName: \"kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222794 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222830 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222854 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggcx7\" (UniqueName: \"kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.222889 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.223345 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.223534 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.223621 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.276118 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78j8\" (UniqueName: \"kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8\") pod \"dnsmasq-dns-6bc7876d45-jl98t\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.324663 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.324741 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.324773 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.324811 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggcx7\" (UniqueName: \"kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.324841 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.325620 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.326177 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.326700 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.327282 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.341717 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggcx7\" (UniqueName: \"kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7\") pod \"dnsmasq-dns-8554648995-47plp\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.514765 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.600482 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.625972 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.636277 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc\") pod \"c695bfa7-8a98-45d4-97df-56955479f855\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.636503 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb\") pod \"c695bfa7-8a98-45d4-97df-56955479f855\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.636601 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config\") pod \"c695bfa7-8a98-45d4-97df-56955479f855\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.636640 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f78j8\" (UniqueName: \"kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8\") pod \"c695bfa7-8a98-45d4-97df-56955479f855\" (UID: \"c695bfa7-8a98-45d4-97df-56955479f855\") " Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.636763 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c695bfa7-8a98-45d4-97df-56955479f855" (UID: "c695bfa7-8a98-45d4-97df-56955479f855"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.637088 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config" (OuterVolumeSpecName: "config") pod "c695bfa7-8a98-45d4-97df-56955479f855" (UID: "c695bfa7-8a98-45d4-97df-56955479f855"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.637384 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c695bfa7-8a98-45d4-97df-56955479f855" (UID: "c695bfa7-8a98-45d4-97df-56955479f855"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.637303 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.637722 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.646892 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8" (OuterVolumeSpecName: "kube-api-access-f78j8") pod "c695bfa7-8a98-45d4-97df-56955479f855" (UID: "c695bfa7-8a98-45d4-97df-56955479f855"). InnerVolumeSpecName "kube-api-access-f78j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.739690 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c695bfa7-8a98-45d4-97df-56955479f855-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.739737 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f78j8\" (UniqueName: \"kubernetes.io/projected/c695bfa7-8a98-45d4-97df-56955479f855-kube-api-access-f78j8\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.853221 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.906218 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:27 crc kubenswrapper[4831]: I1124 08:31:27.977338 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.033494 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.081952 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.401229 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.401292 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.401351 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.401964 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.402028 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40" gracePeriod=600 Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.645281 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40" exitCode=0 Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.645380 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40"} Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.645416 4831 scope.go:117] "RemoveContainer" containerID="5c1869d0589b332aa7e33b85a96ed139bb1c7a86480cb83d293caecc7401f090" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.650973 4831 generic.go:334] "Generic (PLEG): container finished" podID="802765b2-9623-4cbb-949a-9c04446dffd0" containerID="1b8fa3bb649db213deb7ccefb70efb72e246da20ebdaf6fd1854e146a3b41f07" exitCode=0 Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.651103 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-47plp" event={"ID":"802765b2-9623-4cbb-949a-9c04446dffd0","Type":"ContainerDied","Data":"1b8fa3bb649db213deb7ccefb70efb72e246da20ebdaf6fd1854e146a3b41f07"} Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.651149 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-47plp" event={"ID":"802765b2-9623-4cbb-949a-9c04446dffd0","Type":"ContainerStarted","Data":"ca8e864ac9a9a8e68a2a7450e124c8f6d2e95f2e0565f6701c8d4e72ea843d74"} Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.651290 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-jl98t" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.652044 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.652067 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.726225 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.733616 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.840022 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-jl98t"] Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.850181 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-jl98t"] Nov 24 08:31:28 crc kubenswrapper[4831]: I1124 08:31:28.915668 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c695bfa7-8a98-45d4-97df-56955479f855" path="/var/lib/kubelet/pods/c695bfa7-8a98-45d4-97df-56955479f855/volumes" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.125355 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.139337 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.149256 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-th2kl" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.149577 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.149622 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.150336 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.196799 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.267251 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-config\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.267593 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt79d\" (UniqueName: \"kubernetes.io/projected/901cd1f1-9eda-4434-abd5-d128f1cd2511-kube-api-access-zt79d\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.267710 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-scripts\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.267821 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.267940 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.268073 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.268170 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.369793 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.369881 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.369945 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-config\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.370004 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt79d\" (UniqueName: \"kubernetes.io/projected/901cd1f1-9eda-4434-abd5-d128f1cd2511-kube-api-access-zt79d\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.370044 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-scripts\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.370076 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.370100 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.371636 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.372192 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-scripts\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.372337 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901cd1f1-9eda-4434-abd5-d128f1cd2511-config\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.388132 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.390141 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.392114 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/901cd1f1-9eda-4434-abd5-d128f1cd2511-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.431214 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt79d\" (UniqueName: \"kubernetes.io/projected/901cd1f1-9eda-4434-abd5-d128f1cd2511-kube-api-access-zt79d\") pod \"ovn-northd-0\" (UID: \"901cd1f1-9eda-4434-abd5-d128f1cd2511\") " pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.551828 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.681400 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899"} Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.698515 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-47plp" event={"ID":"802765b2-9623-4cbb-949a-9c04446dffd0","Type":"ContainerStarted","Data":"3ba4a3ec89f5016354761a2640f7a0c6338795895ff716e40ec3cb693d12d062"} Nov 24 08:31:29 crc kubenswrapper[4831]: I1124 08:31:29.722904 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-47plp" podStartSLOduration=2.722881553 podStartE2EDuration="2.722881553s" podCreationTimestamp="2025-11-24 08:31:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:31:29.716499801 +0000 UTC m=+963.591644954" watchObservedRunningTime="2025-11-24 08:31:29.722881553 +0000 UTC m=+963.598026706" Nov 24 08:31:30 crc kubenswrapper[4831]: I1124 08:31:30.020893 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 24 08:31:30 crc kubenswrapper[4831]: W1124 08:31:30.028701 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod901cd1f1_9eda_4434_abd5_d128f1cd2511.slice/crio-acf424e5d8eea57562d24152c40b797e8cd5636866b9457768ce411fa0fc7c8a WatchSource:0}: Error finding container acf424e5d8eea57562d24152c40b797e8cd5636866b9457768ce411fa0fc7c8a: Status 404 returned error can't find the container with id acf424e5d8eea57562d24152c40b797e8cd5636866b9457768ce411fa0fc7c8a Nov 24 08:31:30 crc kubenswrapper[4831]: I1124 08:31:30.705460 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"901cd1f1-9eda-4434-abd5-d128f1cd2511","Type":"ContainerStarted","Data":"acf424e5d8eea57562d24152c40b797e8cd5636866b9457768ce411fa0fc7c8a"} Nov 24 08:31:30 crc kubenswrapper[4831]: I1124 08:31:30.706742 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:31 crc kubenswrapper[4831]: I1124 08:31:31.714745 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"901cd1f1-9eda-4434-abd5-d128f1cd2511","Type":"ContainerStarted","Data":"90cd7565cc12956cebc2c5369c95588befa68f786dd0841a49a083f05e822c0a"} Nov 24 08:31:31 crc kubenswrapper[4831]: I1124 08:31:31.716466 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"901cd1f1-9eda-4434-abd5-d128f1cd2511","Type":"ContainerStarted","Data":"ba699a013c5701b96a71a238ad39e71bd0eaa6c08394841ce1ab192f171c623e"} Nov 24 08:31:31 crc kubenswrapper[4831]: I1124 08:31:31.927782 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.836627838 podStartE2EDuration="2.927759049s" podCreationTimestamp="2025-11-24 08:31:29 +0000 UTC" firstStartedPulling="2025-11-24 08:31:30.030737555 +0000 UTC m=+963.905882708" lastFinishedPulling="2025-11-24 08:31:31.121868766 +0000 UTC m=+964.997013919" observedRunningTime="2025-11-24 08:31:31.738588529 +0000 UTC m=+965.613733692" watchObservedRunningTime="2025-11-24 08:31:31.927759049 +0000 UTC m=+965.802904202" Nov 24 08:31:32 crc kubenswrapper[4831]: I1124 08:31:32.725622 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3294a544-5049-47a8-8633-2250f41005ea","Type":"ContainerStarted","Data":"cd2098e6b3f498379b7b998f5714d4e29a6764be559af2a2f95b22ad2487511a"} Nov 24 08:31:32 crc kubenswrapper[4831]: I1124 08:31:32.725873 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 24 08:31:32 crc kubenswrapper[4831]: I1124 08:31:32.725932 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 08:31:32 crc kubenswrapper[4831]: I1124 08:31:32.746635 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.592465012 podStartE2EDuration="36.746604281s" podCreationTimestamp="2025-11-24 08:30:56 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.179191747 +0000 UTC m=+941.054336900" lastFinishedPulling="2025-11-24 08:31:32.333331016 +0000 UTC m=+966.208476169" observedRunningTime="2025-11-24 08:31:32.738016267 +0000 UTC m=+966.613161450" watchObservedRunningTime="2025-11-24 08:31:32.746604281 +0000 UTC m=+966.621749464" Nov 24 08:31:33 crc kubenswrapper[4831]: I1124 08:31:33.579521 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 24 08:31:33 crc kubenswrapper[4831]: I1124 08:31:33.579729 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 24 08:31:33 crc kubenswrapper[4831]: I1124 08:31:33.662627 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 24 08:31:33 crc kubenswrapper[4831]: I1124 08:31:33.792244 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.711728 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.711785 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.784448 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.853439 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.913037 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-rh8vp"] Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.914284 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:34 crc kubenswrapper[4831]: I1124 08:31:34.924567 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rh8vp"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.017966 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7923-account-create-tdcf5"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.019258 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.023772 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.027478 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7923-account-create-tdcf5"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.074973 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmcvd\" (UniqueName: \"kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.075169 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.176423 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94j88\" (UniqueName: \"kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.177105 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.177228 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmcvd\" (UniqueName: \"kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.177366 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.178031 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.193555 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-c8rqp"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.194865 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.202206 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c8rqp"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.212617 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmcvd\" (UniqueName: \"kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd\") pod \"keystone-db-create-rh8vp\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.236272 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.279092 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7rcw\" (UniqueName: \"kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.279152 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.279184 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.279227 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94j88\" (UniqueName: \"kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.280354 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.306586 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94j88\" (UniqueName: \"kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88\") pod \"keystone-7923-account-create-tdcf5\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.318537 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-qdhq6"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.329846 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.330587 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qdhq6"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.338142 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.342400 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-eb35-account-create-958fz"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.343695 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.348481 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.371945 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-eb35-account-create-958fz"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.381396 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.381505 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbwsc\" (UniqueName: \"kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.381627 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7rcw\" (UniqueName: \"kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.381894 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.381958 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.382727 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.382784 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvjw8\" (UniqueName: \"kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.420039 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7rcw\" (UniqueName: \"kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw\") pod \"placement-db-create-c8rqp\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.456311 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-99fa-account-create-vvht6"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.457686 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.460735 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.463912 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-99fa-account-create-vvht6"] Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.483881 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkq25\" (UniqueName: \"kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.483990 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbwsc\" (UniqueName: \"kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.484044 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.484070 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.484110 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvjw8\" (UniqueName: \"kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.484263 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.485218 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.486721 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.527921 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbwsc\" (UniqueName: \"kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc\") pod \"placement-eb35-account-create-958fz\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.530448 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvjw8\" (UniqueName: \"kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8\") pod \"glance-db-create-qdhq6\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.559548 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.585736 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.585825 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkq25\" (UniqueName: \"kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.586942 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.603433 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkq25\" (UniqueName: \"kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25\") pod \"glance-99fa-account-create-vvht6\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.749248 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.767624 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.800059 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.833237 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-rh8vp"] Nov 24 08:31:35 crc kubenswrapper[4831]: W1124 08:31:35.842043 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d7f359_ef83_48f5_8d48_8256acbd0045.slice/crio-a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b WatchSource:0}: Error finding container a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b: Status 404 returned error can't find the container with id a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b Nov 24 08:31:35 crc kubenswrapper[4831]: I1124 08:31:35.967299 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7923-account-create-tdcf5"] Nov 24 08:31:35 crc kubenswrapper[4831]: W1124 08:31:35.993133 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d3545d9_9a47_49a5_8353_751829dfc507.slice/crio-39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922 WatchSource:0}: Error finding container 39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922: Status 404 returned error can't find the container with id 39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.081734 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-c8rqp"] Nov 24 08:31:36 crc kubenswrapper[4831]: W1124 08:31:36.101297 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2312deb6_6491_4cd3_8e15_967215e9d9e0.slice/crio-00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4 WatchSource:0}: Error finding container 00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4: Status 404 returned error can't find the container with id 00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.311351 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-qdhq6"] Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.406353 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-eb35-account-create-958fz"] Nov 24 08:31:36 crc kubenswrapper[4831]: W1124 08:31:36.410104 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod496f9648_9656_4eff_9589_03132e32cec7.slice/crio-fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe WatchSource:0}: Error finding container fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe: Status 404 returned error can't find the container with id fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.414427 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-99fa-account-create-vvht6"] Nov 24 08:31:36 crc kubenswrapper[4831]: W1124 08:31:36.419183 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podece974ef_9247_4703_8277_65eab2031685.slice/crio-c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153 WatchSource:0}: Error finding container c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153: Status 404 returned error can't find the container with id c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.789415 4831 generic.go:334] "Generic (PLEG): container finished" podID="6d3545d9-9a47-49a5-8353-751829dfc507" containerID="4efc22cca0bede96ca0107647c613a9ce44818b36ecb45bd03b404b1d678ff7b" exitCode=0 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.789653 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7923-account-create-tdcf5" event={"ID":"6d3545d9-9a47-49a5-8353-751829dfc507","Type":"ContainerDied","Data":"4efc22cca0bede96ca0107647c613a9ce44818b36ecb45bd03b404b1d678ff7b"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.789698 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7923-account-create-tdcf5" event={"ID":"6d3545d9-9a47-49a5-8353-751829dfc507","Type":"ContainerStarted","Data":"39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.803286 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99fa-account-create-vvht6" event={"ID":"ece974ef-9247-4703-8277-65eab2031685","Type":"ContainerStarted","Data":"f4162bfa538d2eade51040e40a35de2e44d974b064e72d756044e40738b20e9e"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.803516 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99fa-account-create-vvht6" event={"ID":"ece974ef-9247-4703-8277-65eab2031685","Type":"ContainerStarted","Data":"c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.808229 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qdhq6" event={"ID":"dc257041-e2dc-415c-8988-53de7db6a961","Type":"ContainerStarted","Data":"28b6539b0a0e0ffb0c03bd8f5b3dd6ad1b5d70635a50d3b364f817c67ed5092d"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.809238 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qdhq6" event={"ID":"dc257041-e2dc-415c-8988-53de7db6a961","Type":"ContainerStarted","Data":"0adf07e0baa8f66fbd3437e1cdf546cce0727d7918cb5ee19595ea3d9d88bf27"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.810765 4831 generic.go:334] "Generic (PLEG): container finished" podID="35d7f359-ef83-48f5-8d48-8256acbd0045" containerID="44977e20af8ef71481775a5bb73eede1e929fadc5ece316252ed28125e636ecc" exitCode=0 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.810917 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rh8vp" event={"ID":"35d7f359-ef83-48f5-8d48-8256acbd0045","Type":"ContainerDied","Data":"44977e20af8ef71481775a5bb73eede1e929fadc5ece316252ed28125e636ecc"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.811013 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rh8vp" event={"ID":"35d7f359-ef83-48f5-8d48-8256acbd0045","Type":"ContainerStarted","Data":"a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.813140 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb35-account-create-958fz" event={"ID":"496f9648-9656-4eff-9589-03132e32cec7","Type":"ContainerStarted","Data":"c54af51b868f996359c36821fc7a715427cda6412e4a41061c91f6143d5ffc58"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.813234 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb35-account-create-958fz" event={"ID":"496f9648-9656-4eff-9589-03132e32cec7","Type":"ContainerStarted","Data":"fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.816214 4831 generic.go:334] "Generic (PLEG): container finished" podID="2312deb6-6491-4cd3-8e15-967215e9d9e0" containerID="f3d9aeeacec416686d8f484558c2f7cda865a83318ff596b6ecb71abcc20cdba" exitCode=0 Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.816374 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8rqp" event={"ID":"2312deb6-6491-4cd3-8e15-967215e9d9e0","Type":"ContainerDied","Data":"f3d9aeeacec416686d8f484558c2f7cda865a83318ff596b6ecb71abcc20cdba"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.816449 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8rqp" event={"ID":"2312deb6-6491-4cd3-8e15-967215e9d9e0","Type":"ContainerStarted","Data":"00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4"} Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.824851 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-99fa-account-create-vvht6" podStartSLOduration=1.824833545 podStartE2EDuration="1.824833545s" podCreationTimestamp="2025-11-24 08:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:31:36.824510666 +0000 UTC m=+970.699655829" watchObservedRunningTime="2025-11-24 08:31:36.824833545 +0000 UTC m=+970.699978698" Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.865996 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-eb35-account-create-958fz" podStartSLOduration=1.8659671169999998 podStartE2EDuration="1.865967117s" podCreationTimestamp="2025-11-24 08:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:31:36.848927452 +0000 UTC m=+970.724072615" watchObservedRunningTime="2025-11-24 08:31:36.865967117 +0000 UTC m=+970.741112270" Nov 24 08:31:36 crc kubenswrapper[4831]: I1124 08:31:36.885894 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-qdhq6" podStartSLOduration=1.885874314 podStartE2EDuration="1.885874314s" podCreationTimestamp="2025-11-24 08:31:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:31:36.883097515 +0000 UTC m=+970.758242668" watchObservedRunningTime="2025-11-24 08:31:36.885874314 +0000 UTC m=+970.761019467" Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.195288 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.516692 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.593286 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.593923 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="dnsmasq-dns" containerID="cri-o://f74ef8d260f27be9f62d68e1bcf97dc381e49fc59caf9472619ef4334ef75252" gracePeriod=10 Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.832388 4831 generic.go:334] "Generic (PLEG): container finished" podID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerID="f74ef8d260f27be9f62d68e1bcf97dc381e49fc59caf9472619ef4334ef75252" exitCode=0 Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.832466 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" event={"ID":"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604","Type":"ContainerDied","Data":"f74ef8d260f27be9f62d68e1bcf97dc381e49fc59caf9472619ef4334ef75252"} Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.834057 4831 generic.go:334] "Generic (PLEG): container finished" podID="ece974ef-9247-4703-8277-65eab2031685" containerID="f4162bfa538d2eade51040e40a35de2e44d974b064e72d756044e40738b20e9e" exitCode=0 Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.834176 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99fa-account-create-vvht6" event={"ID":"ece974ef-9247-4703-8277-65eab2031685","Type":"ContainerDied","Data":"f4162bfa538d2eade51040e40a35de2e44d974b064e72d756044e40738b20e9e"} Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.837296 4831 generic.go:334] "Generic (PLEG): container finished" podID="dc257041-e2dc-415c-8988-53de7db6a961" containerID="28b6539b0a0e0ffb0c03bd8f5b3dd6ad1b5d70635a50d3b364f817c67ed5092d" exitCode=0 Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.837345 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qdhq6" event={"ID":"dc257041-e2dc-415c-8988-53de7db6a961","Type":"ContainerDied","Data":"28b6539b0a0e0ffb0c03bd8f5b3dd6ad1b5d70635a50d3b364f817c67ed5092d"} Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.839801 4831 generic.go:334] "Generic (PLEG): container finished" podID="496f9648-9656-4eff-9589-03132e32cec7" containerID="c54af51b868f996359c36821fc7a715427cda6412e4a41061c91f6143d5ffc58" exitCode=0 Nov 24 08:31:37 crc kubenswrapper[4831]: I1124 08:31:37.839839 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb35-account-create-958fz" event={"ID":"496f9648-9656-4eff-9589-03132e32cec7","Type":"ContainerDied","Data":"c54af51b868f996359c36821fc7a715427cda6412e4a41061c91f6143d5ffc58"} Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.090981 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.238066 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config\") pod \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.238205 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc\") pod \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.238259 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97ks9\" (UniqueName: \"kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9\") pod \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\" (UID: \"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.260107 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9" (OuterVolumeSpecName: "kube-api-access-97ks9") pod "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" (UID: "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604"). InnerVolumeSpecName "kube-api-access-97ks9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.340972 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97ks9\" (UniqueName: \"kubernetes.io/projected/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-kube-api-access-97ks9\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.344995 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config" (OuterVolumeSpecName: "config") pod "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" (UID: "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.350116 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" (UID: "b91ad2e2-a4e7-4cbe-a06f-cf37720b5604"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.442663 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.442715 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.468243 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.476462 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.482963 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.543677 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7rcw\" (UniqueName: \"kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw\") pod \"2312deb6-6491-4cd3-8e15-967215e9d9e0\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.543733 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts\") pod \"2312deb6-6491-4cd3-8e15-967215e9d9e0\" (UID: \"2312deb6-6491-4cd3-8e15-967215e9d9e0\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.544494 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2312deb6-6491-4cd3-8e15-967215e9d9e0" (UID: "2312deb6-6491-4cd3-8e15-967215e9d9e0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.547030 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw" (OuterVolumeSpecName: "kube-api-access-q7rcw") pod "2312deb6-6491-4cd3-8e15-967215e9d9e0" (UID: "2312deb6-6491-4cd3-8e15-967215e9d9e0"). InnerVolumeSpecName "kube-api-access-q7rcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.645436 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts\") pod \"35d7f359-ef83-48f5-8d48-8256acbd0045\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.645776 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94j88\" (UniqueName: \"kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88\") pod \"6d3545d9-9a47-49a5-8353-751829dfc507\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.645827 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts\") pod \"6d3545d9-9a47-49a5-8353-751829dfc507\" (UID: \"6d3545d9-9a47-49a5-8353-751829dfc507\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.645855 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35d7f359-ef83-48f5-8d48-8256acbd0045" (UID: "35d7f359-ef83-48f5-8d48-8256acbd0045"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.645880 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmcvd\" (UniqueName: \"kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd\") pod \"35d7f359-ef83-48f5-8d48-8256acbd0045\" (UID: \"35d7f359-ef83-48f5-8d48-8256acbd0045\") " Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.646210 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7rcw\" (UniqueName: \"kubernetes.io/projected/2312deb6-6491-4cd3-8e15-967215e9d9e0-kube-api-access-q7rcw\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.646224 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2312deb6-6491-4cd3-8e15-967215e9d9e0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.646231 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d7f359-ef83-48f5-8d48-8256acbd0045-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.646518 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d3545d9-9a47-49a5-8353-751829dfc507" (UID: "6d3545d9-9a47-49a5-8353-751829dfc507"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.648757 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88" (OuterVolumeSpecName: "kube-api-access-94j88") pod "6d3545d9-9a47-49a5-8353-751829dfc507" (UID: "6d3545d9-9a47-49a5-8353-751829dfc507"). InnerVolumeSpecName "kube-api-access-94j88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.649226 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd" (OuterVolumeSpecName: "kube-api-access-zmcvd") pod "35d7f359-ef83-48f5-8d48-8256acbd0045" (UID: "35d7f359-ef83-48f5-8d48-8256acbd0045"). InnerVolumeSpecName "kube-api-access-zmcvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.747851 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d3545d9-9a47-49a5-8353-751829dfc507-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.748112 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmcvd\" (UniqueName: \"kubernetes.io/projected/35d7f359-ef83-48f5-8d48-8256acbd0045-kube-api-access-zmcvd\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.748235 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94j88\" (UniqueName: \"kubernetes.io/projected/6d3545d9-9a47-49a5-8353-751829dfc507-kube-api-access-94j88\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.851484 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-rh8vp" event={"ID":"35d7f359-ef83-48f5-8d48-8256acbd0045","Type":"ContainerDied","Data":"a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b"} Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.851519 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6d99cb0dafc9ab91349f81989e0f3f15d858d869699293b32c3eb90faf3233b" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.851611 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-rh8vp" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.853422 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" event={"ID":"b91ad2e2-a4e7-4cbe-a06f-cf37720b5604","Type":"ContainerDied","Data":"d80b866f7c15f10f303fda9e154916f04fdcb86472d0fd078a4038708f86a079"} Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.853471 4831 scope.go:117] "RemoveContainer" containerID="f74ef8d260f27be9f62d68e1bcf97dc381e49fc59caf9472619ef4334ef75252" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.853661 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9slcr" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.856539 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-c8rqp" event={"ID":"2312deb6-6491-4cd3-8e15-967215e9d9e0","Type":"ContainerDied","Data":"00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4"} Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.856568 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00c55d9f64a51216ada25d8a1aee6d021930e7c007a728eebe92effe49589fd4" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.856569 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-c8rqp" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.859109 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7923-account-create-tdcf5" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.859545 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7923-account-create-tdcf5" event={"ID":"6d3545d9-9a47-49a5-8353-751829dfc507","Type":"ContainerDied","Data":"39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922"} Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.859571 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39d0356523d8bf168d0543651d82c6bb79e4ce509dc1884fb60e1fbadb4fc922" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.890481 4831 scope.go:117] "RemoveContainer" containerID="aad9024288c938941c0bef90bfd7d139de74a98c106f686d56127d941841b3d5" Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.966911 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:31:38 crc kubenswrapper[4831]: I1124 08:31:38.973260 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9slcr"] Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.356012 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.465188 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvjw8\" (UniqueName: \"kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8\") pod \"dc257041-e2dc-415c-8988-53de7db6a961\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.465351 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts\") pod \"dc257041-e2dc-415c-8988-53de7db6a961\" (UID: \"dc257041-e2dc-415c-8988-53de7db6a961\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.466075 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc257041-e2dc-415c-8988-53de7db6a961" (UID: "dc257041-e2dc-415c-8988-53de7db6a961"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.470626 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8" (OuterVolumeSpecName: "kube-api-access-xvjw8") pod "dc257041-e2dc-415c-8988-53de7db6a961" (UID: "dc257041-e2dc-415c-8988-53de7db6a961"). InnerVolumeSpecName "kube-api-access-xvjw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.534279 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.540659 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.567380 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc257041-e2dc-415c-8988-53de7db6a961-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.567412 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvjw8\" (UniqueName: \"kubernetes.io/projected/dc257041-e2dc-415c-8988-53de7db6a961-kube-api-access-xvjw8\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.668831 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkq25\" (UniqueName: \"kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25\") pod \"ece974ef-9247-4703-8277-65eab2031685\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.668926 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts\") pod \"ece974ef-9247-4703-8277-65eab2031685\" (UID: \"ece974ef-9247-4703-8277-65eab2031685\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.669006 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbwsc\" (UniqueName: \"kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc\") pod \"496f9648-9656-4eff-9589-03132e32cec7\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.669172 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts\") pod \"496f9648-9656-4eff-9589-03132e32cec7\" (UID: \"496f9648-9656-4eff-9589-03132e32cec7\") " Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.669481 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ece974ef-9247-4703-8277-65eab2031685" (UID: "ece974ef-9247-4703-8277-65eab2031685"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.669713 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "496f9648-9656-4eff-9589-03132e32cec7" (UID: "496f9648-9656-4eff-9589-03132e32cec7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.671823 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc" (OuterVolumeSpecName: "kube-api-access-gbwsc") pod "496f9648-9656-4eff-9589-03132e32cec7" (UID: "496f9648-9656-4eff-9589-03132e32cec7"). InnerVolumeSpecName "kube-api-access-gbwsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.672122 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25" (OuterVolumeSpecName: "kube-api-access-fkq25") pod "ece974ef-9247-4703-8277-65eab2031685" (UID: "ece974ef-9247-4703-8277-65eab2031685"). InnerVolumeSpecName "kube-api-access-fkq25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.771061 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkq25\" (UniqueName: \"kubernetes.io/projected/ece974ef-9247-4703-8277-65eab2031685-kube-api-access-fkq25\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.771106 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ece974ef-9247-4703-8277-65eab2031685-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.771145 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbwsc\" (UniqueName: \"kubernetes.io/projected/496f9648-9656-4eff-9589-03132e32cec7-kube-api-access-gbwsc\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.771154 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/496f9648-9656-4eff-9589-03132e32cec7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.867838 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-qdhq6" event={"ID":"dc257041-e2dc-415c-8988-53de7db6a961","Type":"ContainerDied","Data":"0adf07e0baa8f66fbd3437e1cdf546cce0727d7918cb5ee19595ea3d9d88bf27"} Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.867886 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0adf07e0baa8f66fbd3437e1cdf546cce0727d7918cb5ee19595ea3d9d88bf27" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.867950 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-qdhq6" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.871239 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb35-account-create-958fz" event={"ID":"496f9648-9656-4eff-9589-03132e32cec7","Type":"ContainerDied","Data":"fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe"} Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.871287 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa137d91861b2fe74ddb2e9d5a69c89f25e50fb92801a6d308a176408dcc4dfe" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.871408 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb35-account-create-958fz" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.886063 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-99fa-account-create-vvht6" event={"ID":"ece974ef-9247-4703-8277-65eab2031685","Type":"ContainerDied","Data":"c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153"} Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.886113 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c41772bcd80ad1457c24eda5cb3e6683961bdcfec4e5c9d279c5dfbf97ef7153" Nov 24 08:31:39 crc kubenswrapper[4831]: I1124 08:31:39.886124 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-99fa-account-create-vvht6" Nov 24 08:31:40 crc kubenswrapper[4831]: I1124 08:31:40.906152 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" path="/var/lib/kubelet/pods/b91ad2e2-a4e7-4cbe-a06f-cf37720b5604/volumes" Nov 24 08:31:44 crc kubenswrapper[4831]: I1124 08:31:44.610922 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.425800 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qzcrh"] Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426431 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d7f359-ef83-48f5-8d48-8256acbd0045" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426447 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d7f359-ef83-48f5-8d48-8256acbd0045" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426473 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d3545d9-9a47-49a5-8353-751829dfc507" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426481 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d3545d9-9a47-49a5-8353-751829dfc507" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426495 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2312deb6-6491-4cd3-8e15-967215e9d9e0" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426505 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="2312deb6-6491-4cd3-8e15-967215e9d9e0" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426519 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496f9648-9656-4eff-9589-03132e32cec7" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426528 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="496f9648-9656-4eff-9589-03132e32cec7" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426539 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="dnsmasq-dns" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426549 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="dnsmasq-dns" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426563 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc257041-e2dc-415c-8988-53de7db6a961" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426572 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc257041-e2dc-415c-8988-53de7db6a961" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426589 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="init" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426599 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="init" Nov 24 08:31:45 crc kubenswrapper[4831]: E1124 08:31:45.426611 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ece974ef-9247-4703-8277-65eab2031685" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426619 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ece974ef-9247-4703-8277-65eab2031685" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426825 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="496f9648-9656-4eff-9589-03132e32cec7" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426858 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b91ad2e2-a4e7-4cbe-a06f-cf37720b5604" containerName="dnsmasq-dns" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426873 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d3545d9-9a47-49a5-8353-751829dfc507" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426893 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="2312deb6-6491-4cd3-8e15-967215e9d9e0" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426921 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc257041-e2dc-415c-8988-53de7db6a961" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426942 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="ece974ef-9247-4703-8277-65eab2031685" containerName="mariadb-account-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.426972 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d7f359-ef83-48f5-8d48-8256acbd0045" containerName="mariadb-database-create" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.427574 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.429589 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.430419 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-slfbn" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.447504 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qzcrh"] Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.580685 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.580761 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.580793 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.580882 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnr4l\" (UniqueName: \"kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.682908 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.682979 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.683009 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.683058 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnr4l\" (UniqueName: \"kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.690107 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.690694 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.691075 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.701344 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnr4l\" (UniqueName: \"kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l\") pod \"glance-db-sync-qzcrh\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:45 crc kubenswrapper[4831]: I1124 08:31:45.751859 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qzcrh" Nov 24 08:31:46 crc kubenswrapper[4831]: I1124 08:31:46.095933 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qzcrh"] Nov 24 08:31:46 crc kubenswrapper[4831]: W1124 08:31:46.102696 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c9b9615_df1a_4680_8460_c150476edd94.slice/crio-1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980 WatchSource:0}: Error finding container 1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980: Status 404 returned error can't find the container with id 1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980 Nov 24 08:31:46 crc kubenswrapper[4831]: I1124 08:31:46.964757 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qzcrh" event={"ID":"0c9b9615-df1a-4680-8460-c150476edd94","Type":"ContainerStarted","Data":"1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980"} Nov 24 08:31:52 crc kubenswrapper[4831]: I1124 08:31:52.238917 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p87rh" podUID="71dd0aae-5932-4334-9639-49a91209160c" containerName="ovn-controller" probeResult="failure" output=< Nov 24 08:31:52 crc kubenswrapper[4831]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 08:31:52 crc kubenswrapper[4831]: > Nov 24 08:31:52 crc kubenswrapper[4831]: I1124 08:31:52.322294 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:54 crc kubenswrapper[4831]: I1124 08:31:54.019580 4831 generic.go:334] "Generic (PLEG): container finished" podID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerID="351e96e33d34f1dd7a1d56c166ca6a8cefdb3058b5fba117be0e7a2f7ffb73d5" exitCode=0 Nov 24 08:31:54 crc kubenswrapper[4831]: I1124 08:31:54.019891 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerDied","Data":"351e96e33d34f1dd7a1d56c166ca6a8cefdb3058b5fba117be0e7a2f7ffb73d5"} Nov 24 08:31:54 crc kubenswrapper[4831]: I1124 08:31:54.026932 4831 generic.go:334] "Generic (PLEG): container finished" podID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerID="6a1f253dccd2eceb4b96368b8a5e1b2330dfc1c9a7d3318435c861f05c2d1f8c" exitCode=0 Nov 24 08:31:54 crc kubenswrapper[4831]: I1124 08:31:54.026973 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerDied","Data":"6a1f253dccd2eceb4b96368b8a5e1b2330dfc1c9a7d3318435c861f05c2d1f8c"} Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.254349 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-p87rh" podUID="71dd0aae-5932-4334-9639-49a91209160c" containerName="ovn-controller" probeResult="failure" output=< Nov 24 08:31:57 crc kubenswrapper[4831]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 24 08:31:57 crc kubenswrapper[4831]: > Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.358140 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-4wbwz" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.569126 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p87rh-config-scg5w"] Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.571620 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.590249 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.604846 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh-config-scg5w"] Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661178 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq9zz\" (UniqueName: \"kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661248 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661269 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661292 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661334 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.661369 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.770797 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq9zz\" (UniqueName: \"kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.770931 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.770976 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.771141 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.771222 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.771350 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.771551 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.771559 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.772575 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.773445 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.775591 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.791152 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq9zz\" (UniqueName: \"kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz\") pod \"ovn-controller-p87rh-config-scg5w\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:57 crc kubenswrapper[4831]: I1124 08:31:57.903688 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.428681 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh-config-scg5w"] Nov 24 08:31:58 crc kubenswrapper[4831]: W1124 08:31:58.435228 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41c83e09_6c80_459e_8ba6_7fce8297ebeb.slice/crio-d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265 WatchSource:0}: Error finding container d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265: Status 404 returned error can't find the container with id d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265 Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.635559 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerStarted","Data":"474b53112447ea9aedc0b761e2bae117da65c20019e0c99f838485074bc9d4d9"} Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.635973 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.637643 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerStarted","Data":"7c831df1604165a09ea1c5798d73a1ef09c377057c5c164dd214cbd9b440ed18"} Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.638167 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.639843 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-scg5w" event={"ID":"41c83e09-6c80-459e-8ba6-7fce8297ebeb","Type":"ContainerStarted","Data":"d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265"} Nov 24 08:31:58 crc kubenswrapper[4831]: I1124 08:31:58.685823 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.702488924 podStartE2EDuration="1m8.685797888s" podCreationTimestamp="2025-11-24 08:30:50 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.470465417 +0000 UTC m=+941.345610570" lastFinishedPulling="2025-11-24 08:31:20.453774381 +0000 UTC m=+954.328919534" observedRunningTime="2025-11-24 08:31:58.682817693 +0000 UTC m=+992.557962856" watchObservedRunningTime="2025-11-24 08:31:58.685797888 +0000 UTC m=+992.560943051" Nov 24 08:31:59 crc kubenswrapper[4831]: I1124 08:31:59.652574 4831 generic.go:334] "Generic (PLEG): container finished" podID="41c83e09-6c80-459e-8ba6-7fce8297ebeb" containerID="46eee5fa1cbc845a0637d7078d8ffb82868b46d45a3de566cb29ba954b534461" exitCode=0 Nov 24 08:31:59 crc kubenswrapper[4831]: I1124 08:31:59.652826 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-scg5w" event={"ID":"41c83e09-6c80-459e-8ba6-7fce8297ebeb","Type":"ContainerDied","Data":"46eee5fa1cbc845a0637d7078d8ffb82868b46d45a3de566cb29ba954b534461"} Nov 24 08:31:59 crc kubenswrapper[4831]: I1124 08:31:59.659647 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qzcrh" event={"ID":"0c9b9615-df1a-4680-8460-c150476edd94","Type":"ContainerStarted","Data":"6c3729170ea6d4c10fe9a03adf4076424673d06e6c77c517119f65470b208a2c"} Nov 24 08:31:59 crc kubenswrapper[4831]: I1124 08:31:59.672613 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.463218185 podStartE2EDuration="1m9.672597896s" podCreationTimestamp="2025-11-24 08:30:50 +0000 UTC" firstStartedPulling="2025-11-24 08:31:07.44986016 +0000 UTC m=+941.325005313" lastFinishedPulling="2025-11-24 08:31:19.659239871 +0000 UTC m=+953.534385024" observedRunningTime="2025-11-24 08:31:58.739844418 +0000 UTC m=+992.614989581" watchObservedRunningTime="2025-11-24 08:31:59.672597896 +0000 UTC m=+993.547743049" Nov 24 08:31:59 crc kubenswrapper[4831]: I1124 08:31:59.695594 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qzcrh" podStartSLOduration=2.930084237 podStartE2EDuration="14.695577371s" podCreationTimestamp="2025-11-24 08:31:45 +0000 UTC" firstStartedPulling="2025-11-24 08:31:46.104733036 +0000 UTC m=+979.979878189" lastFinishedPulling="2025-11-24 08:31:57.87022617 +0000 UTC m=+991.745371323" observedRunningTime="2025-11-24 08:31:59.690395663 +0000 UTC m=+993.565540826" watchObservedRunningTime="2025-11-24 08:31:59.695577371 +0000 UTC m=+993.570722524" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.095498 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.234691 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.234782 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.234875 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.234981 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq9zz\" (UniqueName: \"kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.235016 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.235037 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn\") pod \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\" (UID: \"41c83e09-6c80-459e-8ba6-7fce8297ebeb\") " Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.235537 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.236005 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run" (OuterVolumeSpecName: "var-run") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.236547 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.237265 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts" (OuterVolumeSpecName: "scripts") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.237506 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.245603 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz" (OuterVolumeSpecName: "kube-api-access-kq9zz") pod "41c83e09-6c80-459e-8ba6-7fce8297ebeb" (UID: "41c83e09-6c80-459e-8ba6-7fce8297ebeb"). InnerVolumeSpecName "kube-api-access-kq9zz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337632 4831 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337685 4831 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337699 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41c83e09-6c80-459e-8ba6-7fce8297ebeb-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337708 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq9zz\" (UniqueName: \"kubernetes.io/projected/41c83e09-6c80-459e-8ba6-7fce8297ebeb-kube-api-access-kq9zz\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337720 4831 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.337731 4831 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/41c83e09-6c80-459e-8ba6-7fce8297ebeb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.690618 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-scg5w" event={"ID":"41c83e09-6c80-459e-8ba6-7fce8297ebeb","Type":"ContainerDied","Data":"d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265"} Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.690692 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d37501701bb2af469813464a6dbb5d89972b4fda020089ced55f10799e89b265" Nov 24 08:32:01 crc kubenswrapper[4831]: I1124 08:32:01.690744 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-scg5w" Nov 24 08:32:01 crc kubenswrapper[4831]: E1124 08:32:01.823988 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41c83e09_6c80_459e_8ba6_7fce8297ebeb.slice\": RecentStats: unable to find data in memory cache]" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.262746 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p87rh-config-scg5w"] Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.277000 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p87rh-config-scg5w"] Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.300220 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-p87rh" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.385484 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-p87rh-config-qxhpd"] Nov 24 08:32:02 crc kubenswrapper[4831]: E1124 08:32:02.386175 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c83e09-6c80-459e-8ba6-7fce8297ebeb" containerName="ovn-config" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.386204 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c83e09-6c80-459e-8ba6-7fce8297ebeb" containerName="ovn-config" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.386488 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c83e09-6c80-459e-8ba6-7fce8297ebeb" containerName="ovn-config" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.387424 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.391853 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.405382 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh-config-qxhpd"] Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456404 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456486 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zcn9\" (UniqueName: \"kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456520 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456538 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456555 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.456610 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557758 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zcn9\" (UniqueName: \"kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557816 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557834 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557851 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557906 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.557947 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.558728 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.559082 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.559109 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.559181 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.565985 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.588957 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zcn9\" (UniqueName: \"kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9\") pod \"ovn-controller-p87rh-config-qxhpd\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.711048 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:02 crc kubenswrapper[4831]: I1124 08:32:02.909985 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c83e09-6c80-459e-8ba6-7fce8297ebeb" path="/var/lib/kubelet/pods/41c83e09-6c80-459e-8ba6-7fce8297ebeb/volumes" Nov 24 08:32:03 crc kubenswrapper[4831]: I1124 08:32:03.010621 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-p87rh-config-qxhpd"] Nov 24 08:32:03 crc kubenswrapper[4831]: I1124 08:32:03.708097 4831 generic.go:334] "Generic (PLEG): container finished" podID="02fde1a9-9db4-403b-b92f-5cc66e7d2b80" containerID="60f3f66a9016ea5a56cf062f8cb3b4f53c6051dbf6b63cf2e933cbb34852e775" exitCode=0 Nov 24 08:32:03 crc kubenswrapper[4831]: I1124 08:32:03.708203 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-qxhpd" event={"ID":"02fde1a9-9db4-403b-b92f-5cc66e7d2b80","Type":"ContainerDied","Data":"60f3f66a9016ea5a56cf062f8cb3b4f53c6051dbf6b63cf2e933cbb34852e775"} Nov 24 08:32:03 crc kubenswrapper[4831]: I1124 08:32:03.708450 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-qxhpd" event={"ID":"02fde1a9-9db4-403b-b92f-5cc66e7d2b80","Type":"ContainerStarted","Data":"07c7e9d199ad0fd77e12d87f806cb96e22b55b36e392e886bb8cbb81dbafe524"} Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.025711 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110736 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110786 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110851 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zcn9\" (UniqueName: \"kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110944 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110959 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.110935 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run" (OuterVolumeSpecName: "var-run") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.111002 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn\") pod \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\" (UID: \"02fde1a9-9db4-403b-b92f-5cc66e7d2b80\") " Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.111034 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.111302 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.111985 4831 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.112036 4831 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.112059 4831 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-var-run\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.111994 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.112390 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts" (OuterVolumeSpecName: "scripts") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.117591 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9" (OuterVolumeSpecName: "kube-api-access-2zcn9") pod "02fde1a9-9db4-403b-b92f-5cc66e7d2b80" (UID: "02fde1a9-9db4-403b-b92f-5cc66e7d2b80"). InnerVolumeSpecName "kube-api-access-2zcn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.213376 4831 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.213826 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zcn9\" (UniqueName: \"kubernetes.io/projected/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-kube-api-access-2zcn9\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.213891 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/02fde1a9-9db4-403b-b92f-5cc66e7d2b80-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.724011 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-p87rh-config-qxhpd" event={"ID":"02fde1a9-9db4-403b-b92f-5cc66e7d2b80","Type":"ContainerDied","Data":"07c7e9d199ad0fd77e12d87f806cb96e22b55b36e392e886bb8cbb81dbafe524"} Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.724273 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07c7e9d199ad0fd77e12d87f806cb96e22b55b36e392e886bb8cbb81dbafe524" Nov 24 08:32:05 crc kubenswrapper[4831]: I1124 08:32:05.724061 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-p87rh-config-qxhpd" Nov 24 08:32:06 crc kubenswrapper[4831]: I1124 08:32:06.113895 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-p87rh-config-qxhpd"] Nov 24 08:32:06 crc kubenswrapper[4831]: I1124 08:32:06.119860 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-p87rh-config-qxhpd"] Nov 24 08:32:06 crc kubenswrapper[4831]: I1124 08:32:06.902721 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02fde1a9-9db4-403b-b92f-5cc66e7d2b80" path="/var/lib/kubelet/pods/02fde1a9-9db4-403b-b92f-5cc66e7d2b80/volumes" Nov 24 08:32:08 crc kubenswrapper[4831]: I1124 08:32:08.747648 4831 generic.go:334] "Generic (PLEG): container finished" podID="0c9b9615-df1a-4680-8460-c150476edd94" containerID="6c3729170ea6d4c10fe9a03adf4076424673d06e6c77c517119f65470b208a2c" exitCode=0 Nov 24 08:32:08 crc kubenswrapper[4831]: I1124 08:32:08.747695 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qzcrh" event={"ID":"0c9b9615-df1a-4680-8460-c150476edd94","Type":"ContainerDied","Data":"6c3729170ea6d4c10fe9a03adf4076424673d06e6c77c517119f65470b208a2c"} Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.137581 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qzcrh" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.189148 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle\") pod \"0c9b9615-df1a-4680-8460-c150476edd94\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.189262 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data\") pod \"0c9b9615-df1a-4680-8460-c150476edd94\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.189311 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnr4l\" (UniqueName: \"kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l\") pod \"0c9b9615-df1a-4680-8460-c150476edd94\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.189390 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data\") pod \"0c9b9615-df1a-4680-8460-c150476edd94\" (UID: \"0c9b9615-df1a-4680-8460-c150476edd94\") " Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.197513 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0c9b9615-df1a-4680-8460-c150476edd94" (UID: "0c9b9615-df1a-4680-8460-c150476edd94"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.205598 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l" (OuterVolumeSpecName: "kube-api-access-nnr4l") pod "0c9b9615-df1a-4680-8460-c150476edd94" (UID: "0c9b9615-df1a-4680-8460-c150476edd94"). InnerVolumeSpecName "kube-api-access-nnr4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.213061 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c9b9615-df1a-4680-8460-c150476edd94" (UID: "0c9b9615-df1a-4680-8460-c150476edd94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.235021 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data" (OuterVolumeSpecName: "config-data") pod "0c9b9615-df1a-4680-8460-c150476edd94" (UID: "0c9b9615-df1a-4680-8460-c150476edd94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.291171 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.291201 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.291214 4831 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9b9615-df1a-4680-8460-c150476edd94-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.291223 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnr4l\" (UniqueName: \"kubernetes.io/projected/0c9b9615-df1a-4680-8460-c150476edd94-kube-api-access-nnr4l\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.766862 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qzcrh" event={"ID":"0c9b9615-df1a-4680-8460-c150476edd94","Type":"ContainerDied","Data":"1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980"} Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.766949 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d4104f48ab45ddcbaf1084dd3afbae5fbd018c02e8a3d5ebac310f8927e5980" Nov 24 08:32:10 crc kubenswrapper[4831]: I1124 08:32:10.766907 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qzcrh" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.196624 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:11 crc kubenswrapper[4831]: E1124 08:32:11.196942 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9b9615-df1a-4680-8460-c150476edd94" containerName="glance-db-sync" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.196957 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9b9615-df1a-4680-8460-c150476edd94" containerName="glance-db-sync" Nov 24 08:32:11 crc kubenswrapper[4831]: E1124 08:32:11.196972 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fde1a9-9db4-403b-b92f-5cc66e7d2b80" containerName="ovn-config" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.196978 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fde1a9-9db4-403b-b92f-5cc66e7d2b80" containerName="ovn-config" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.197144 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9b9615-df1a-4680-8460-c150476edd94" containerName="glance-db-sync" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.197168 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fde1a9-9db4-403b-b92f-5cc66e7d2b80" containerName="ovn-config" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.198043 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.224379 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.307050 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.307184 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.307222 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.307264 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.307348 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plfdl\" (UniqueName: \"kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.408600 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.408653 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.408687 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.408731 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plfdl\" (UniqueName: \"kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.408777 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.409822 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.410003 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.410312 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.410304 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.430884 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plfdl\" (UniqueName: \"kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl\") pod \"dnsmasq-dns-554567b4f7-gzw82\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.519618 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.661455 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 08:32:11 crc kubenswrapper[4831]: I1124 08:32:11.973203 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.232822 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.444502 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wk2jg"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.446010 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.483268 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wk2jg"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.533904 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.534224 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qclg7\" (UniqueName: \"kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.625227 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xmkr7"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.626482 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.635571 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.635828 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qclg7\" (UniqueName: \"kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.636784 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.648748 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xmkr7"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.705174 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qclg7\" (UniqueName: \"kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7\") pod \"cinder-db-create-wk2jg\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.737303 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.737398 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vcrv\" (UniqueName: \"kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.760103 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-041b-account-create-xz99f"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.761111 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.767955 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.783538 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" event={"ID":"7053a17c-f398-4bb0-bc6d-30718f4e4596","Type":"ContainerStarted","Data":"f7b1c2ae02daff8b374dd2f523f2de3046f150105f90ce1115303b44ec06b0ed"} Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.798018 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.826138 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-041b-account-create-xz99f"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.843711 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2wss\" (UniqueName: \"kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.844044 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.844239 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.844385 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vcrv\" (UniqueName: \"kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.845037 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.946364 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2wss\" (UniqueName: \"kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.946449 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.947393 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.963367 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-daa7-account-create-rbtz7"] Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.968122 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.973218 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vcrv\" (UniqueName: \"kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv\") pod \"barbican-db-create-xmkr7\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.988883 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 24 08:32:12 crc kubenswrapper[4831]: I1124 08:32:12.995727 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-daa7-account-create-rbtz7"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.003035 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2wss\" (UniqueName: \"kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss\") pod \"barbican-041b-account-create-xz99f\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.049206 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8l52\" (UniqueName: \"kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.049618 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.087692 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.133528 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-frfnc"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.134955 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.146809 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-frfnc"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.151557 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8l52\" (UniqueName: \"kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.151668 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.152450 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.187827 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-m95zj"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.188960 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.193198 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.193722 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m2wp6" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.193939 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.201130 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.242062 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.254185 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpzqh\" (UniqueName: \"kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.254270 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.254292 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.254355 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.254376 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rshjf\" (UniqueName: \"kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.259239 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m95zj"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.259735 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8l52\" (UniqueName: \"kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52\") pod \"cinder-daa7-account-create-rbtz7\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.328596 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a381-account-create-5jxnk"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.334082 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.349685 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358477 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpzqh\" (UniqueName: \"kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358539 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6cfl\" (UniqueName: \"kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358573 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358597 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358653 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358681 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rshjf\" (UniqueName: \"kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.358705 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.368883 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.379376 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a381-account-create-5jxnk"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.385187 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.385399 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.407010 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.418277 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpzqh\" (UniqueName: \"kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh\") pod \"neutron-db-create-frfnc\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.425837 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rshjf\" (UniqueName: \"kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf\") pod \"keystone-db-sync-m95zj\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.463601 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.469836 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6cfl\" (UniqueName: \"kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.469968 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.470735 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.517121 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6cfl\" (UniqueName: \"kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl\") pod \"neutron-a381-account-create-5jxnk\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.519087 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.690283 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.820251 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-041b-account-create-xz99f"] Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.843627 4831 generic.go:334] "Generic (PLEG): container finished" podID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerID="c6a70ab3e6e7fd5de6bddd955bb8ac2a7fd90703d7bd33db94b924a3a91638c9" exitCode=0 Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.843670 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" event={"ID":"7053a17c-f398-4bb0-bc6d-30718f4e4596","Type":"ContainerDied","Data":"c6a70ab3e6e7fd5de6bddd955bb8ac2a7fd90703d7bd33db94b924a3a91638c9"} Nov 24 08:32:13 crc kubenswrapper[4831]: W1124 08:32:13.857019 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f3b2d59_b9aa_44ae_a7ca_2f597a08435e.slice/crio-ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe WatchSource:0}: Error finding container ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe: Status 404 returned error can't find the container with id ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe Nov 24 08:32:13 crc kubenswrapper[4831]: I1124 08:32:13.991057 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wk2jg"] Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.223980 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xmkr7"] Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.462534 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-daa7-account-create-rbtz7"] Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.480995 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-frfnc"] Nov 24 08:32:14 crc kubenswrapper[4831]: W1124 08:32:14.505087 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc01f8f81_1e14_40a3_a1ed_195d691d9fba.slice/crio-69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653 WatchSource:0}: Error finding container 69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653: Status 404 returned error can't find the container with id 69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653 Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.654678 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a381-account-create-5jxnk"] Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.793736 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-m95zj"] Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.876819 4831 generic.go:334] "Generic (PLEG): container finished" podID="2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" containerID="03e5863499760678fbd083899acc363f4824eff8f205f899dec1e2724e5ab6db" exitCode=0 Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.877453 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-041b-account-create-xz99f" event={"ID":"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e","Type":"ContainerDied","Data":"03e5863499760678fbd083899acc363f4824eff8f205f899dec1e2724e5ab6db"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.877518 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-041b-account-create-xz99f" event={"ID":"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e","Type":"ContainerStarted","Data":"ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.882877 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frfnc" event={"ID":"c01f8f81-1e14-40a3-a1ed-195d691d9fba","Type":"ContainerStarted","Data":"69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.912995 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m95zj" event={"ID":"843cd907-38e7-401e-ac60-69a7c3b17610","Type":"ContainerStarted","Data":"82d61ac4f153a79898d7e0f2528f698a9a0e65aa69aed928effc7ffc2c06d841"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.913054 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-daa7-account-create-rbtz7" event={"ID":"3e80d4e0-3649-4a08-b1ec-d606b6be82eb","Type":"ContainerStarted","Data":"597bb417f3675dfc91fdaa83e848185a35b3ba5c86fe451ce7d8bb6dc38c4bfb"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.922778 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" event={"ID":"7053a17c-f398-4bb0-bc6d-30718f4e4596","Type":"ContainerStarted","Data":"34d0cf823622d256971a1a5d35b13a955f183879d1ee63b4a474d32aaa2eba64"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.924375 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.930537 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wk2jg" event={"ID":"89bc8072-5ead-41da-9dec-790ab8cd3d58","Type":"ContainerStarted","Data":"34862a0c923d434d7c056566cd18b3c11d7da2a82dd454d85ca2aa6a7d1f696d"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.930579 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wk2jg" event={"ID":"89bc8072-5ead-41da-9dec-790ab8cd3d58","Type":"ContainerStarted","Data":"75ae1afef9bcd31982d6bf922d6e5060b27cc7523806cfb0365cd2a37f22ff8d"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.936760 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xmkr7" event={"ID":"dd0e715a-72c6-4240-9dac-fd3efb994bff","Type":"ContainerStarted","Data":"97fb9d8197a30d388f67688d72bd6be91d48f09824d8d96b2f68a14a31c2a3e8"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.936802 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xmkr7" event={"ID":"dd0e715a-72c6-4240-9dac-fd3efb994bff","Type":"ContainerStarted","Data":"1cc1401a6f62ee2ae439061e2c098c66cdfd630f326248e2bfa272f64fb35aee"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.940298 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a381-account-create-5jxnk" event={"ID":"01fbdeba-6c1f-423e-92b0-6e69136756f1","Type":"ContainerStarted","Data":"1b2443fe6d5d51366cf1781b79c9d90def4ad42ec614151d5c31d9836d4c349e"} Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.955524 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" podStartSLOduration=3.9555049650000003 podStartE2EDuration="3.955504965s" podCreationTimestamp="2025-11-24 08:32:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:14.948867096 +0000 UTC m=+1008.824012249" watchObservedRunningTime="2025-11-24 08:32:14.955504965 +0000 UTC m=+1008.830650118" Nov 24 08:32:14 crc kubenswrapper[4831]: I1124 08:32:14.968654 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-wk2jg" podStartSLOduration=2.968636489 podStartE2EDuration="2.968636489s" podCreationTimestamp="2025-11-24 08:32:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:14.965665145 +0000 UTC m=+1008.840810318" watchObservedRunningTime="2025-11-24 08:32:14.968636489 +0000 UTC m=+1008.843781642" Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.970722 4831 generic.go:334] "Generic (PLEG): container finished" podID="dd0e715a-72c6-4240-9dac-fd3efb994bff" containerID="97fb9d8197a30d388f67688d72bd6be91d48f09824d8d96b2f68a14a31c2a3e8" exitCode=0 Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.971416 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xmkr7" event={"ID":"dd0e715a-72c6-4240-9dac-fd3efb994bff","Type":"ContainerDied","Data":"97fb9d8197a30d388f67688d72bd6be91d48f09824d8d96b2f68a14a31c2a3e8"} Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.982435 4831 generic.go:334] "Generic (PLEG): container finished" podID="01fbdeba-6c1f-423e-92b0-6e69136756f1" containerID="8e1699bd4161a3cf314a2d2b35a8698dfc0a30ba026922b7233279171af08b08" exitCode=0 Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.982508 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a381-account-create-5jxnk" event={"ID":"01fbdeba-6c1f-423e-92b0-6e69136756f1","Type":"ContainerDied","Data":"8e1699bd4161a3cf314a2d2b35a8698dfc0a30ba026922b7233279171af08b08"} Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.991920 4831 generic.go:334] "Generic (PLEG): container finished" podID="c01f8f81-1e14-40a3-a1ed-195d691d9fba" containerID="6058f3c92d28aabd97db5819d0be4d04e680ffcd16cf93e85d461a3d10c1f653" exitCode=0 Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.991988 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frfnc" event={"ID":"c01f8f81-1e14-40a3-a1ed-195d691d9fba","Type":"ContainerDied","Data":"6058f3c92d28aabd97db5819d0be4d04e680ffcd16cf93e85d461a3d10c1f653"} Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.994247 4831 generic.go:334] "Generic (PLEG): container finished" podID="3e80d4e0-3649-4a08-b1ec-d606b6be82eb" containerID="024e47611b52191ac038ef064f4a8addb1e41eafdcea32506c90c2ddbab87816" exitCode=0 Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.994307 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-daa7-account-create-rbtz7" event={"ID":"3e80d4e0-3649-4a08-b1ec-d606b6be82eb","Type":"ContainerDied","Data":"024e47611b52191ac038ef064f4a8addb1e41eafdcea32506c90c2ddbab87816"} Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.996016 4831 generic.go:334] "Generic (PLEG): container finished" podID="89bc8072-5ead-41da-9dec-790ab8cd3d58" containerID="34862a0c923d434d7c056566cd18b3c11d7da2a82dd454d85ca2aa6a7d1f696d" exitCode=0 Nov 24 08:32:15 crc kubenswrapper[4831]: I1124 08:32:15.996071 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wk2jg" event={"ID":"89bc8072-5ead-41da-9dec-790ab8cd3d58","Type":"ContainerDied","Data":"34862a0c923d434d7c056566cd18b3c11d7da2a82dd454d85ca2aa6a7d1f696d"} Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.487465 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.589619 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2wss\" (UniqueName: \"kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss\") pod \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.589715 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts\") pod \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\" (UID: \"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e\") " Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.590981 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" (UID: "2f3b2d59-b9aa-44ae-a7ca-2f597a08435e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.591205 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.596578 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss" (OuterVolumeSpecName: "kube-api-access-h2wss") pod "2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" (UID: "2f3b2d59-b9aa-44ae-a7ca-2f597a08435e"). InnerVolumeSpecName "kube-api-access-h2wss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:16 crc kubenswrapper[4831]: I1124 08:32:16.693628 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2wss\" (UniqueName: \"kubernetes.io/projected/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e-kube-api-access-h2wss\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.006022 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-041b-account-create-xz99f" event={"ID":"2f3b2d59-b9aa-44ae-a7ca-2f597a08435e","Type":"ContainerDied","Data":"ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe"} Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.006720 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-041b-account-create-xz99f" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.007352 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecf4b51291c3ea3193e2a55f3f5a1b42bfce839e6151ba3cdf4153ceef215afe" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.463597 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.609214 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8l52\" (UniqueName: \"kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52\") pod \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.611745 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts\") pod \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\" (UID: \"3e80d4e0-3649-4a08-b1ec-d606b6be82eb\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.612472 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e80d4e0-3649-4a08-b1ec-d606b6be82eb" (UID: "3e80d4e0-3649-4a08-b1ec-d606b6be82eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.634078 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52" (OuterVolumeSpecName: "kube-api-access-r8l52") pod "3e80d4e0-3649-4a08-b1ec-d606b6be82eb" (UID: "3e80d4e0-3649-4a08-b1ec-d606b6be82eb"). InnerVolumeSpecName "kube-api-access-r8l52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.713978 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.714024 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8l52\" (UniqueName: \"kubernetes.io/projected/3e80d4e0-3649-4a08-b1ec-d606b6be82eb-kube-api-access-r8l52\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.742808 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.750067 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.751308 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.770040 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815343 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpzqh\" (UniqueName: \"kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh\") pod \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815429 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts\") pod \"dd0e715a-72c6-4240-9dac-fd3efb994bff\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815497 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts\") pod \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\" (UID: \"c01f8f81-1e14-40a3-a1ed-195d691d9fba\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815559 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts\") pod \"89bc8072-5ead-41da-9dec-790ab8cd3d58\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815597 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qclg7\" (UniqueName: \"kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7\") pod \"89bc8072-5ead-41da-9dec-790ab8cd3d58\" (UID: \"89bc8072-5ead-41da-9dec-790ab8cd3d58\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.815633 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vcrv\" (UniqueName: \"kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv\") pod \"dd0e715a-72c6-4240-9dac-fd3efb994bff\" (UID: \"dd0e715a-72c6-4240-9dac-fd3efb994bff\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.817927 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c01f8f81-1e14-40a3-a1ed-195d691d9fba" (UID: "c01f8f81-1e14-40a3-a1ed-195d691d9fba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.818298 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dd0e715a-72c6-4240-9dac-fd3efb994bff" (UID: "dd0e715a-72c6-4240-9dac-fd3efb994bff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.818685 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "89bc8072-5ead-41da-9dec-790ab8cd3d58" (UID: "89bc8072-5ead-41da-9dec-790ab8cd3d58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.820729 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv" (OuterVolumeSpecName: "kube-api-access-5vcrv") pod "dd0e715a-72c6-4240-9dac-fd3efb994bff" (UID: "dd0e715a-72c6-4240-9dac-fd3efb994bff"). InnerVolumeSpecName "kube-api-access-5vcrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.823109 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7" (OuterVolumeSpecName: "kube-api-access-qclg7") pod "89bc8072-5ead-41da-9dec-790ab8cd3d58" (UID: "89bc8072-5ead-41da-9dec-790ab8cd3d58"). InnerVolumeSpecName "kube-api-access-qclg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.835661 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh" (OuterVolumeSpecName: "kube-api-access-rpzqh") pod "c01f8f81-1e14-40a3-a1ed-195d691d9fba" (UID: "c01f8f81-1e14-40a3-a1ed-195d691d9fba"). InnerVolumeSpecName "kube-api-access-rpzqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917504 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6cfl\" (UniqueName: \"kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl\") pod \"01fbdeba-6c1f-423e-92b0-6e69136756f1\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917565 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts\") pod \"01fbdeba-6c1f-423e-92b0-6e69136756f1\" (UID: \"01fbdeba-6c1f-423e-92b0-6e69136756f1\") " Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917904 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dd0e715a-72c6-4240-9dac-fd3efb994bff-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917920 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01f8f81-1e14-40a3-a1ed-195d691d9fba-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917930 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/89bc8072-5ead-41da-9dec-790ab8cd3d58-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917939 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qclg7\" (UniqueName: \"kubernetes.io/projected/89bc8072-5ead-41da-9dec-790ab8cd3d58-kube-api-access-qclg7\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917949 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vcrv\" (UniqueName: \"kubernetes.io/projected/dd0e715a-72c6-4240-9dac-fd3efb994bff-kube-api-access-5vcrv\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.917958 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpzqh\" (UniqueName: \"kubernetes.io/projected/c01f8f81-1e14-40a3-a1ed-195d691d9fba-kube-api-access-rpzqh\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.918081 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01fbdeba-6c1f-423e-92b0-6e69136756f1" (UID: "01fbdeba-6c1f-423e-92b0-6e69136756f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:17 crc kubenswrapper[4831]: I1124 08:32:17.925610 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl" (OuterVolumeSpecName: "kube-api-access-q6cfl") pod "01fbdeba-6c1f-423e-92b0-6e69136756f1" (UID: "01fbdeba-6c1f-423e-92b0-6e69136756f1"). InnerVolumeSpecName "kube-api-access-q6cfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.015559 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xmkr7" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.015573 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xmkr7" event={"ID":"dd0e715a-72c6-4240-9dac-fd3efb994bff","Type":"ContainerDied","Data":"1cc1401a6f62ee2ae439061e2c098c66cdfd630f326248e2bfa272f64fb35aee"} Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.015630 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cc1401a6f62ee2ae439061e2c098c66cdfd630f326248e2bfa272f64fb35aee" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.020030 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6cfl\" (UniqueName: \"kubernetes.io/projected/01fbdeba-6c1f-423e-92b0-6e69136756f1-kube-api-access-q6cfl\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.020252 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01fbdeba-6c1f-423e-92b0-6e69136756f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.028059 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a381-account-create-5jxnk" event={"ID":"01fbdeba-6c1f-423e-92b0-6e69136756f1","Type":"ContainerDied","Data":"1b2443fe6d5d51366cf1781b79c9d90def4ad42ec614151d5c31d9836d4c349e"} Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.028094 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b2443fe6d5d51366cf1781b79c9d90def4ad42ec614151d5c31d9836d4c349e" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.028154 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a381-account-create-5jxnk" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.035441 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-frfnc" event={"ID":"c01f8f81-1e14-40a3-a1ed-195d691d9fba","Type":"ContainerDied","Data":"69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653"} Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.035658 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69a54fb3aed746d96c2465511b27e9a6e791f87098fe3141e968ac47dca73653" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.035559 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-frfnc" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.044349 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-daa7-account-create-rbtz7" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.044340 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-daa7-account-create-rbtz7" event={"ID":"3e80d4e0-3649-4a08-b1ec-d606b6be82eb","Type":"ContainerDied","Data":"597bb417f3675dfc91fdaa83e848185a35b3ba5c86fe451ce7d8bb6dc38c4bfb"} Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.045817 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="597bb417f3675dfc91fdaa83e848185a35b3ba5c86fe451ce7d8bb6dc38c4bfb" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.047111 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wk2jg" event={"ID":"89bc8072-5ead-41da-9dec-790ab8cd3d58","Type":"ContainerDied","Data":"75ae1afef9bcd31982d6bf922d6e5060b27cc7523806cfb0365cd2a37f22ff8d"} Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.047141 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75ae1afef9bcd31982d6bf922d6e5060b27cc7523806cfb0365cd2a37f22ff8d" Nov 24 08:32:18 crc kubenswrapper[4831]: I1124 08:32:18.047197 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wk2jg" Nov 24 08:32:21 crc kubenswrapper[4831]: I1124 08:32:21.520981 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:21 crc kubenswrapper[4831]: I1124 08:32:21.572915 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:32:21 crc kubenswrapper[4831]: I1124 08:32:21.573213 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-47plp" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="dnsmasq-dns" containerID="cri-o://3ba4a3ec89f5016354761a2640f7a0c6338795895ff716e40ec3cb693d12d062" gracePeriod=10 Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.102387 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m95zj" event={"ID":"843cd907-38e7-401e-ac60-69a7c3b17610","Type":"ContainerStarted","Data":"622430eb11da2ae1d25bcc31862089471717187021823e69cbfce98b462c515f"} Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.117762 4831 generic.go:334] "Generic (PLEG): container finished" podID="802765b2-9623-4cbb-949a-9c04446dffd0" containerID="3ba4a3ec89f5016354761a2640f7a0c6338795895ff716e40ec3cb693d12d062" exitCode=0 Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.117853 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-47plp" event={"ID":"802765b2-9623-4cbb-949a-9c04446dffd0","Type":"ContainerDied","Data":"3ba4a3ec89f5016354761a2640f7a0c6338795895ff716e40ec3cb693d12d062"} Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.117919 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-47plp" event={"ID":"802765b2-9623-4cbb-949a-9c04446dffd0","Type":"ContainerDied","Data":"ca8e864ac9a9a8e68a2a7450e124c8f6d2e95f2e0565f6701c8d4e72ea843d74"} Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.117935 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca8e864ac9a9a8e68a2a7450e124c8f6d2e95f2e0565f6701c8d4e72ea843d74" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.133613 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-m95zj" podStartSLOduration=2.6605488360000003 podStartE2EDuration="9.133591637s" podCreationTimestamp="2025-11-24 08:32:13 +0000 UTC" firstStartedPulling="2025-11-24 08:32:14.807492178 +0000 UTC m=+1008.682637331" lastFinishedPulling="2025-11-24 08:32:21.280534979 +0000 UTC m=+1015.155680132" observedRunningTime="2025-11-24 08:32:22.129595253 +0000 UTC m=+1016.004740426" watchObservedRunningTime="2025-11-24 08:32:22.133591637 +0000 UTC m=+1016.008736810" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.141739 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.248344 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc\") pod \"802765b2-9623-4cbb-949a-9c04446dffd0\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.248411 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config\") pod \"802765b2-9623-4cbb-949a-9c04446dffd0\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.248465 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb\") pod \"802765b2-9623-4cbb-949a-9c04446dffd0\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.248534 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggcx7\" (UniqueName: \"kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7\") pod \"802765b2-9623-4cbb-949a-9c04446dffd0\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.248585 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb\") pod \"802765b2-9623-4cbb-949a-9c04446dffd0\" (UID: \"802765b2-9623-4cbb-949a-9c04446dffd0\") " Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.253887 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7" (OuterVolumeSpecName: "kube-api-access-ggcx7") pod "802765b2-9623-4cbb-949a-9c04446dffd0" (UID: "802765b2-9623-4cbb-949a-9c04446dffd0"). InnerVolumeSpecName "kube-api-access-ggcx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.312268 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "802765b2-9623-4cbb-949a-9c04446dffd0" (UID: "802765b2-9623-4cbb-949a-9c04446dffd0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.327735 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "802765b2-9623-4cbb-949a-9c04446dffd0" (UID: "802765b2-9623-4cbb-949a-9c04446dffd0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.328432 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "802765b2-9623-4cbb-949a-9c04446dffd0" (UID: "802765b2-9623-4cbb-949a-9c04446dffd0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.332625 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config" (OuterVolumeSpecName: "config") pod "802765b2-9623-4cbb-949a-9c04446dffd0" (UID: "802765b2-9623-4cbb-949a-9c04446dffd0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.352715 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.353362 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.353451 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.353589 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggcx7\" (UniqueName: \"kubernetes.io/projected/802765b2-9623-4cbb-949a-9c04446dffd0-kube-api-access-ggcx7\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:22 crc kubenswrapper[4831]: I1124 08:32:22.353677 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/802765b2-9623-4cbb-949a-9c04446dffd0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:23 crc kubenswrapper[4831]: I1124 08:32:23.124500 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-47plp" Nov 24 08:32:23 crc kubenswrapper[4831]: I1124 08:32:23.152713 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:32:23 crc kubenswrapper[4831]: I1124 08:32:23.163208 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-47plp"] Nov 24 08:32:24 crc kubenswrapper[4831]: I1124 08:32:24.902198 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" path="/var/lib/kubelet/pods/802765b2-9623-4cbb-949a-9c04446dffd0/volumes" Nov 24 08:32:25 crc kubenswrapper[4831]: I1124 08:32:25.142083 4831 generic.go:334] "Generic (PLEG): container finished" podID="843cd907-38e7-401e-ac60-69a7c3b17610" containerID="622430eb11da2ae1d25bcc31862089471717187021823e69cbfce98b462c515f" exitCode=0 Nov 24 08:32:25 crc kubenswrapper[4831]: I1124 08:32:25.142129 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m95zj" event={"ID":"843cd907-38e7-401e-ac60-69a7c3b17610","Type":"ContainerDied","Data":"622430eb11da2ae1d25bcc31862089471717187021823e69cbfce98b462c515f"} Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.541578 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.724726 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rshjf\" (UniqueName: \"kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf\") pod \"843cd907-38e7-401e-ac60-69a7c3b17610\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.724849 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle\") pod \"843cd907-38e7-401e-ac60-69a7c3b17610\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.724957 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data\") pod \"843cd907-38e7-401e-ac60-69a7c3b17610\" (UID: \"843cd907-38e7-401e-ac60-69a7c3b17610\") " Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.737172 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf" (OuterVolumeSpecName: "kube-api-access-rshjf") pod "843cd907-38e7-401e-ac60-69a7c3b17610" (UID: "843cd907-38e7-401e-ac60-69a7c3b17610"). InnerVolumeSpecName "kube-api-access-rshjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.750789 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "843cd907-38e7-401e-ac60-69a7c3b17610" (UID: "843cd907-38e7-401e-ac60-69a7c3b17610"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.770378 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data" (OuterVolumeSpecName: "config-data") pod "843cd907-38e7-401e-ac60-69a7c3b17610" (UID: "843cd907-38e7-401e-ac60-69a7c3b17610"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.828282 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rshjf\" (UniqueName: \"kubernetes.io/projected/843cd907-38e7-401e-ac60-69a7c3b17610-kube-api-access-rshjf\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.828397 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:26 crc kubenswrapper[4831]: I1124 08:32:26.828424 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/843cd907-38e7-401e-ac60-69a7c3b17610-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.169166 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-m95zj" event={"ID":"843cd907-38e7-401e-ac60-69a7c3b17610","Type":"ContainerDied","Data":"82d61ac4f153a79898d7e0f2528f698a9a0e65aa69aed928effc7ffc2c06d841"} Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.169637 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82d61ac4f153a79898d7e0f2528f698a9a0e65aa69aed928effc7ffc2c06d841" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.169770 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-m95zj" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.447790 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448238 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="843cd907-38e7-401e-ac60-69a7c3b17610" containerName="keystone-db-sync" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448266 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="843cd907-38e7-401e-ac60-69a7c3b17610" containerName="keystone-db-sync" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448290 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="init" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448300 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="init" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448337 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0e715a-72c6-4240-9dac-fd3efb994bff" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448347 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0e715a-72c6-4240-9dac-fd3efb994bff" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448366 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fbdeba-6c1f-423e-92b0-6e69136756f1" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448373 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fbdeba-6c1f-423e-92b0-6e69136756f1" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448387 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bc8072-5ead-41da-9dec-790ab8cd3d58" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448395 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bc8072-5ead-41da-9dec-790ab8cd3d58" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448407 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e80d4e0-3649-4a08-b1ec-d606b6be82eb" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448415 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e80d4e0-3649-4a08-b1ec-d606b6be82eb" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448427 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01f8f81-1e14-40a3-a1ed-195d691d9fba" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448434 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01f8f81-1e14-40a3-a1ed-195d691d9fba" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448447 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448455 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: E1124 08:32:27.448474 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="dnsmasq-dns" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448482 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="dnsmasq-dns" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448674 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0e715a-72c6-4240-9dac-fd3efb994bff" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448697 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e80d4e0-3649-4a08-b1ec-d606b6be82eb" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448715 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="01fbdeba-6c1f-423e-92b0-6e69136756f1" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448731 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01f8f81-1e14-40a3-a1ed-195d691d9fba" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448744 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" containerName="mariadb-account-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448768 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="89bc8072-5ead-41da-9dec-790ab8cd3d58" containerName="mariadb-database-create" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448781 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="843cd907-38e7-401e-ac60-69a7c3b17610" containerName="keystone-db-sync" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.448794 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="802765b2-9623-4cbb-949a-9c04446dffd0" containerName="dnsmasq-dns" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.449909 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.464046 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.538761 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-txvnh"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.542516 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.558337 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.558604 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m2wp6" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.558758 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.558942 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.559603 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.575277 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-txvnh"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.642225 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.642276 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.642330 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.642363 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6bm\" (UniqueName: \"kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.642378 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744506 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6bm\" (UniqueName: \"kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744553 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744581 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744619 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744675 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744706 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744731 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744756 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744779 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744803 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zjhv\" (UniqueName: \"kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.744825 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.745657 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.746121 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.746395 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.746860 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.774247 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6bm\" (UniqueName: \"kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm\") pod \"dnsmasq-dns-67795cd9-xvf2b\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.833785 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.836694 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.844830 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.845409 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846310 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846443 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846477 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846535 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zjhv\" (UniqueName: \"kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846576 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.846627 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.875976 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.887009 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.891664 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.896520 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.917133 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.943030 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zjhv\" (UniqueName: \"kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949306 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq2d9\" (UniqueName: \"kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949373 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949393 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949440 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949474 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949512 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.949534 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.956040 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data\") pod \"keystone-bootstrap-txvnh\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.956840 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tqhj9"] Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.958387 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.975103 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 08:32:27 crc kubenswrapper[4831]: I1124 08:32:27.993606 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:27.998494 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-b89k8" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.052532 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tqhj9"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054193 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054223 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054267 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054287 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054352 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054369 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054386 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054430 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054450 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054468 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054510 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054582 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq2d9\" (UniqueName: \"kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.054605 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcvtd\" (UniqueName: \"kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.055787 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.056009 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.096162 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.102063 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.110392 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.118364 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.130181 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.161761 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.161841 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.161898 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcvtd\" (UniqueName: \"kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.161955 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.161991 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.162010 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.171046 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.172801 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.183102 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq2d9\" (UniqueName: \"kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9\") pod \"ceilometer-0\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.184878 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.186665 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.188782 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.189746 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.205189 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcvtd\" (UniqueName: \"kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd\") pod \"cinder-db-sync-tqhj9\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.209416 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qqjkx"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.210710 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.238114 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.238342 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fjjgc" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.287436 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qqjkx"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.305655 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.306334 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.367006 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vm69f"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.368008 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.369926 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9s2s\" (UniqueName: \"kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.370010 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.370077 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.381631 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xzmx2" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.382099 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.382373 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.437385 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vm69f"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.450479 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.474388 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nzfjn"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.475979 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.476101 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.476810 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9s2s\" (UniqueName: \"kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.476844 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.478469 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8p65\" (UniqueName: \"kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.478536 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.478577 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.502110 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.502385 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vjpmw" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.502511 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.516740 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9s2s\" (UniqueName: \"kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.517209 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.537080 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle\") pod \"barbican-db-sync-qqjkx\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.559491 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nzfjn"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581180 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581226 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581261 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8p65\" (UniqueName: \"kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581289 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581328 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581352 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t68ps\" (UniqueName: \"kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.581436 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.589434 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.597161 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.620523 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.634166 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8p65\" (UniqueName: \"kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65\") pod \"neutron-db-sync-vm69f\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.634420 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.689261 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.689365 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t68ps\" (UniqueName: \"kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.689493 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.689672 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.689714 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.690388 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.705663 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.735522 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.748998 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.765026 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.768367 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.793923 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpz7m\" (UniqueName: \"kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.794430 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.794476 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.794493 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.794543 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.810161 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t68ps\" (UniqueName: \"kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps\") pod \"placement-db-sync-nzfjn\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.829838 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.895730 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.895818 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpz7m\" (UniqueName: \"kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.895843 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.895877 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.895890 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.898447 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.901389 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nzfjn" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.902360 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.902651 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.905409 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:28 crc kubenswrapper[4831]: I1124 08:32:28.962436 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpz7m\" (UniqueName: \"kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m\") pod \"dnsmasq-dns-5b6dbdb6f5-jstgx\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.003881 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:29 crc kubenswrapper[4831]: W1124 08:32:29.178490 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod007f7857_a198_4abf_a464_e4ab5cf0531a.slice/crio-c7e1049cca4fccae422a0a1932d986c8712bc8f8e2b37b083e6c48922efb4ab3 WatchSource:0}: Error finding container c7e1049cca4fccae422a0a1932d986c8712bc8f8e2b37b083e6c48922efb4ab3: Status 404 returned error can't find the container with id c7e1049cca4fccae422a0a1932d986c8712bc8f8e2b37b083e6c48922efb4ab3 Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.201726 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.226256 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" event={"ID":"007f7857-a198-4abf-a464-e4ab5cf0531a","Type":"ContainerStarted","Data":"c7e1049cca4fccae422a0a1932d986c8712bc8f8e2b37b083e6c48922efb4ab3"} Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.432828 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tqhj9"] Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.477752 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-txvnh"] Nov 24 08:32:29 crc kubenswrapper[4831]: W1124 08:32:29.484893 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb58d631b_1aab_4c67_9496_3742eecb511b.slice/crio-ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6 WatchSource:0}: Error finding container ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6: Status 404 returned error can't find the container with id ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6 Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.711514 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vm69f"] Nov 24 08:32:29 crc kubenswrapper[4831]: W1124 08:32:29.719064 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6303dec0_5e57_4012_b093_1307fd1211e9.slice/crio-d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6 WatchSource:0}: Error finding container d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6: Status 404 returned error can't find the container with id d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6 Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.764094 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.774932 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qqjkx"] Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.887876 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nzfjn"] Nov 24 08:32:29 crc kubenswrapper[4831]: I1124 08:32:29.901648 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.241739 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerStarted","Data":"78959f21897d720ef11a9f0a8e0739ab784a02f478fc45a2eed9a0c7d60509ee"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.243903 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nzfjn" event={"ID":"b76e6599-c355-4d06-80bd-cb66824a64a4","Type":"ContainerStarted","Data":"bc667bd4399949ecea663531d2497f59bcf1e5eebec622e32ae63b1456d0e78c"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.244819 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qqjkx" event={"ID":"d698dc95-8dfe-483d-82ac-21584ce29659","Type":"ContainerStarted","Data":"7fbf1620f4cb40ff91142f2191ed1702103fdd49c5f5ad8054f561afe1f2602c"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.246845 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-txvnh" event={"ID":"b58d631b-1aab-4c67-9496-3742eecb511b","Type":"ContainerStarted","Data":"0ce1190f12c5cf3014548a6949bc4eb34b5c6220a376f40766c7eaf59f2f90d0"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.246863 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-txvnh" event={"ID":"b58d631b-1aab-4c67-9496-3742eecb511b","Type":"ContainerStarted","Data":"ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.252103 4831 generic.go:334] "Generic (PLEG): container finished" podID="59446fe2-a549-4797-8fbd-370904369dd4" containerID="859aab855c9969fbbbd1cd147ed965f97eb96360b69d06469885cf4f0f695f1c" exitCode=0 Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.252222 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" event={"ID":"59446fe2-a549-4797-8fbd-370904369dd4","Type":"ContainerDied","Data":"859aab855c9969fbbbd1cd147ed965f97eb96360b69d06469885cf4f0f695f1c"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.252267 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" event={"ID":"59446fe2-a549-4797-8fbd-370904369dd4","Type":"ContainerStarted","Data":"26cdb90bc8af95515f1216a382201aac22172459d2c1c86f44b3b20f5e6cbe89"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.255904 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vm69f" event={"ID":"6303dec0-5e57-4012-b093-1307fd1211e9","Type":"ContainerStarted","Data":"0056e5daaedbf3c446a0301f7940251ec01fa21fbba3f93937b2300d5e991173"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.255940 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vm69f" event={"ID":"6303dec0-5e57-4012-b093-1307fd1211e9","Type":"ContainerStarted","Data":"d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.257707 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tqhj9" event={"ID":"46a54561-e390-48c6-a27d-e516a013ebb3","Type":"ContainerStarted","Data":"fb6a7d29b9ebfbef183ef07fb8c8b9816646431e7a3015127564cb3b19c3dc9f"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.277642 4831 generic.go:334] "Generic (PLEG): container finished" podID="007f7857-a198-4abf-a464-e4ab5cf0531a" containerID="107dfa43d9cae1633dc370404deb484388ee1981873b873a421beb06e332bff9" exitCode=0 Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.277708 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" event={"ID":"007f7857-a198-4abf-a464-e4ab5cf0531a","Type":"ContainerDied","Data":"107dfa43d9cae1633dc370404deb484388ee1981873b873a421beb06e332bff9"} Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.280516 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-txvnh" podStartSLOduration=3.280498603 podStartE2EDuration="3.280498603s" podCreationTimestamp="2025-11-24 08:32:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:30.277802987 +0000 UTC m=+1024.152948140" watchObservedRunningTime="2025-11-24 08:32:30.280498603 +0000 UTC m=+1024.155643746" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.307080 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vm69f" podStartSLOduration=2.30705714 podStartE2EDuration="2.30705714s" podCreationTimestamp="2025-11-24 08:32:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:30.302564012 +0000 UTC m=+1024.177709185" watchObservedRunningTime="2025-11-24 08:32:30.30705714 +0000 UTC m=+1024.182202293" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.681566 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.849914 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config\") pod \"007f7857-a198-4abf-a464-e4ab5cf0531a\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.850051 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb\") pod \"007f7857-a198-4abf-a464-e4ab5cf0531a\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.850087 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc\") pod \"007f7857-a198-4abf-a464-e4ab5cf0531a\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.850149 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6bm\" (UniqueName: \"kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm\") pod \"007f7857-a198-4abf-a464-e4ab5cf0531a\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.850247 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb\") pod \"007f7857-a198-4abf-a464-e4ab5cf0531a\" (UID: \"007f7857-a198-4abf-a464-e4ab5cf0531a\") " Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.896214 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm" (OuterVolumeSpecName: "kube-api-access-rw6bm") pod "007f7857-a198-4abf-a464-e4ab5cf0531a" (UID: "007f7857-a198-4abf-a464-e4ab5cf0531a"). InnerVolumeSpecName "kube-api-access-rw6bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.923124 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "007f7857-a198-4abf-a464-e4ab5cf0531a" (UID: "007f7857-a198-4abf-a464-e4ab5cf0531a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.926268 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "007f7857-a198-4abf-a464-e4ab5cf0531a" (UID: "007f7857-a198-4abf-a464-e4ab5cf0531a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.966723 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "007f7857-a198-4abf-a464-e4ab5cf0531a" (UID: "007f7857-a198-4abf-a464-e4ab5cf0531a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.969268 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.969292 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.969308 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6bm\" (UniqueName: \"kubernetes.io/projected/007f7857-a198-4abf-a464-e4ab5cf0531a-kube-api-access-rw6bm\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:30 crc kubenswrapper[4831]: I1124 08:32:30.969332 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.056585 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config" (OuterVolumeSpecName: "config") pod "007f7857-a198-4abf-a464-e4ab5cf0531a" (UID: "007f7857-a198-4abf-a464-e4ab5cf0531a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.071255 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/007f7857-a198-4abf-a464-e4ab5cf0531a-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.096745 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.304685 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" event={"ID":"59446fe2-a549-4797-8fbd-370904369dd4","Type":"ContainerStarted","Data":"2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4"} Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.305602 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.337893 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.338598 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xvf2b" event={"ID":"007f7857-a198-4abf-a464-e4ab5cf0531a","Type":"ContainerDied","Data":"c7e1049cca4fccae422a0a1932d986c8712bc8f8e2b37b083e6c48922efb4ab3"} Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.338695 4831 scope.go:117] "RemoveContainer" containerID="107dfa43d9cae1633dc370404deb484388ee1981873b873a421beb06e332bff9" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.349177 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" podStartSLOduration=3.349141203 podStartE2EDuration="3.349141203s" podCreationTimestamp="2025-11-24 08:32:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:31.331657444 +0000 UTC m=+1025.206802607" watchObservedRunningTime="2025-11-24 08:32:31.349141203 +0000 UTC m=+1025.224286366" Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.482213 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:31 crc kubenswrapper[4831]: I1124 08:32:31.488762 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xvf2b"] Nov 24 08:32:32 crc kubenswrapper[4831]: I1124 08:32:32.908505 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="007f7857-a198-4abf-a464-e4ab5cf0531a" path="/var/lib/kubelet/pods/007f7857-a198-4abf-a464-e4ab5cf0531a/volumes" Nov 24 08:32:35 crc kubenswrapper[4831]: I1124 08:32:35.388255 4831 generic.go:334] "Generic (PLEG): container finished" podID="b58d631b-1aab-4c67-9496-3742eecb511b" containerID="0ce1190f12c5cf3014548a6949bc4eb34b5c6220a376f40766c7eaf59f2f90d0" exitCode=0 Nov 24 08:32:35 crc kubenswrapper[4831]: I1124 08:32:35.388355 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-txvnh" event={"ID":"b58d631b-1aab-4c67-9496-3742eecb511b","Type":"ContainerDied","Data":"0ce1190f12c5cf3014548a6949bc4eb34b5c6220a376f40766c7eaf59f2f90d0"} Nov 24 08:32:36 crc kubenswrapper[4831]: I1124 08:32:36.907554 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010004 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010072 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010097 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010233 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zjhv\" (UniqueName: \"kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010435 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.010491 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data\") pod \"b58d631b-1aab-4c67-9496-3742eecb511b\" (UID: \"b58d631b-1aab-4c67-9496-3742eecb511b\") " Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.055597 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv" (OuterVolumeSpecName: "kube-api-access-7zjhv") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "kube-api-access-7zjhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.055617 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.056580 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts" (OuterVolumeSpecName: "scripts") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.059797 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.067939 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.073071 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data" (OuterVolumeSpecName: "config-data") pod "b58d631b-1aab-4c67-9496-3742eecb511b" (UID: "b58d631b-1aab-4c67-9496-3742eecb511b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112278 4831 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112396 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112409 4831 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112417 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zjhv\" (UniqueName: \"kubernetes.io/projected/b58d631b-1aab-4c67-9496-3742eecb511b-kube-api-access-7zjhv\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112427 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.112434 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b58d631b-1aab-4c67-9496-3742eecb511b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.409560 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-txvnh" event={"ID":"b58d631b-1aab-4c67-9496-3742eecb511b","Type":"ContainerDied","Data":"ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6"} Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.409622 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba3bbfa0c45759dffe1689d0cd5ba81a20cbb56bf39c828e89885612ea1688c6" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.409686 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-txvnh" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.490981 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-txvnh"] Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.501131 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-txvnh"] Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.579128 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tv9wr"] Nov 24 08:32:37 crc kubenswrapper[4831]: E1124 08:32:37.579695 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="007f7857-a198-4abf-a464-e4ab5cf0531a" containerName="init" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.579717 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="007f7857-a198-4abf-a464-e4ab5cf0531a" containerName="init" Nov 24 08:32:37 crc kubenswrapper[4831]: E1124 08:32:37.579744 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b58d631b-1aab-4c67-9496-3742eecb511b" containerName="keystone-bootstrap" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.579753 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b58d631b-1aab-4c67-9496-3742eecb511b" containerName="keystone-bootstrap" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.579990 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="007f7857-a198-4abf-a464-e4ab5cf0531a" containerName="init" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.580032 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b58d631b-1aab-4c67-9496-3742eecb511b" containerName="keystone-bootstrap" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.580838 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.585413 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.585482 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.585642 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.585838 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m2wp6" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.587276 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tv9wr"] Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.588903 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.722630 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.722725 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.722791 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.722868 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.722914 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52h7t\" (UniqueName: \"kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.723012 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.825316 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.826056 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.826102 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.826130 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.826177 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.826203 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52h7t\" (UniqueName: \"kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.832820 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.833845 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.833881 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.838799 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.840226 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.849474 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52h7t\" (UniqueName: \"kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t\") pod \"keystone-bootstrap-tv9wr\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:37 crc kubenswrapper[4831]: I1124 08:32:37.903730 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:32:38 crc kubenswrapper[4831]: I1124 08:32:38.906527 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b58d631b-1aab-4c67-9496-3742eecb511b" path="/var/lib/kubelet/pods/b58d631b-1aab-4c67-9496-3742eecb511b/volumes" Nov 24 08:32:39 crc kubenswrapper[4831]: I1124 08:32:39.005562 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:32:39 crc kubenswrapper[4831]: I1124 08:32:39.071619 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:39 crc kubenswrapper[4831]: I1124 08:32:39.072186 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="dnsmasq-dns" containerID="cri-o://34d0cf823622d256971a1a5d35b13a955f183879d1ee63b4a474d32aaa2eba64" gracePeriod=10 Nov 24 08:32:39 crc kubenswrapper[4831]: I1124 08:32:39.428775 4831 generic.go:334] "Generic (PLEG): container finished" podID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerID="34d0cf823622d256971a1a5d35b13a955f183879d1ee63b4a474d32aaa2eba64" exitCode=0 Nov 24 08:32:39 crc kubenswrapper[4831]: I1124 08:32:39.429054 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" event={"ID":"7053a17c-f398-4bb0-bc6d-30718f4e4596","Type":"ContainerDied","Data":"34d0cf823622d256971a1a5d35b13a955f183879d1ee63b4a474d32aaa2eba64"} Nov 24 08:32:41 crc kubenswrapper[4831]: I1124 08:32:41.520147 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: connect: connection refused" Nov 24 08:32:43 crc kubenswrapper[4831]: E1124 08:32:43.020963 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Nov 24 08:32:43 crc kubenswrapper[4831]: E1124 08:32:43.021587 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5h64dhdh57h658h88h9ch559h6h68dh659h96h55fh75h64h5cch9fh68h554h98h678hbdh5b6h65fh56hcfh6bhfbh574hc8h545hb9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gq2d9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.097833 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.253500 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb\") pod \"7053a17c-f398-4bb0-bc6d-30718f4e4596\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.253926 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config\") pod \"7053a17c-f398-4bb0-bc6d-30718f4e4596\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.253952 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc\") pod \"7053a17c-f398-4bb0-bc6d-30718f4e4596\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.255578 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plfdl\" (UniqueName: \"kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl\") pod \"7053a17c-f398-4bb0-bc6d-30718f4e4596\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.255666 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb\") pod \"7053a17c-f398-4bb0-bc6d-30718f4e4596\" (UID: \"7053a17c-f398-4bb0-bc6d-30718f4e4596\") " Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.259968 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl" (OuterVolumeSpecName: "kube-api-access-plfdl") pod "7053a17c-f398-4bb0-bc6d-30718f4e4596" (UID: "7053a17c-f398-4bb0-bc6d-30718f4e4596"). InnerVolumeSpecName "kube-api-access-plfdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.290262 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7053a17c-f398-4bb0-bc6d-30718f4e4596" (UID: "7053a17c-f398-4bb0-bc6d-30718f4e4596"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.294567 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config" (OuterVolumeSpecName: "config") pod "7053a17c-f398-4bb0-bc6d-30718f4e4596" (UID: "7053a17c-f398-4bb0-bc6d-30718f4e4596"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.299627 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7053a17c-f398-4bb0-bc6d-30718f4e4596" (UID: "7053a17c-f398-4bb0-bc6d-30718f4e4596"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.304754 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7053a17c-f398-4bb0-bc6d-30718f4e4596" (UID: "7053a17c-f398-4bb0-bc6d-30718f4e4596"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.358441 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.358469 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.358479 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.358487 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7053a17c-f398-4bb0-bc6d-30718f4e4596-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.358495 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plfdl\" (UniqueName: \"kubernetes.io/projected/7053a17c-f398-4bb0-bc6d-30718f4e4596-kube-api-access-plfdl\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.485184 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" event={"ID":"7053a17c-f398-4bb0-bc6d-30718f4e4596","Type":"ContainerDied","Data":"f7b1c2ae02daff8b374dd2f523f2de3046f150105f90ce1115303b44ec06b0ed"} Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.485298 4831 scope.go:117] "RemoveContainer" containerID="34d0cf823622d256971a1a5d35b13a955f183879d1ee63b4a474d32aaa2eba64" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.485346 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-gzw82" Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.518112 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:43 crc kubenswrapper[4831]: I1124 08:32:43.530185 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-gzw82"] Nov 24 08:32:44 crc kubenswrapper[4831]: I1124 08:32:44.904455 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" path="/var/lib/kubelet/pods/7053a17c-f398-4bb0-bc6d-30718f4e4596/volumes" Nov 24 08:32:51 crc kubenswrapper[4831]: I1124 08:32:51.573143 4831 generic.go:334] "Generic (PLEG): container finished" podID="6303dec0-5e57-4012-b093-1307fd1211e9" containerID="0056e5daaedbf3c446a0301f7940251ec01fa21fbba3f93937b2300d5e991173" exitCode=0 Nov 24 08:32:51 crc kubenswrapper[4831]: I1124 08:32:51.573579 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vm69f" event={"ID":"6303dec0-5e57-4012-b093-1307fd1211e9","Type":"ContainerDied","Data":"0056e5daaedbf3c446a0301f7940251ec01fa21fbba3f93937b2300d5e991173"} Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.001087 4831 scope.go:117] "RemoveContainer" containerID="c6a70ab3e6e7fd5de6bddd955bb8ac2a7fd90703d7bd33db94b924a3a91638c9" Nov 24 08:32:52 crc kubenswrapper[4831]: E1124 08:32:52.043698 4831 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 24 08:32:52 crc kubenswrapper[4831]: E1124 08:32:52.044285 4831 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pcvtd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-tqhj9_openstack(46a54561-e390-48c6-a27d-e516a013ebb3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 24 08:32:52 crc kubenswrapper[4831]: E1124 08:32:52.045734 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-tqhj9" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.426275 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tv9wr"] Nov 24 08:32:52 crc kubenswrapper[4831]: W1124 08:32:52.438918 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb702363c_0c9b_4804_a478_01ea54210d2c.slice/crio-58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe WatchSource:0}: Error finding container 58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe: Status 404 returned error can't find the container with id 58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.590455 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tv9wr" event={"ID":"b702363c-0c9b-4804-a478-01ea54210d2c","Type":"ContainerStarted","Data":"58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe"} Nov 24 08:32:52 crc kubenswrapper[4831]: E1124 08:32:52.593717 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-tqhj9" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.861275 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.918927 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config\") pod \"6303dec0-5e57-4012-b093-1307fd1211e9\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.919007 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle\") pod \"6303dec0-5e57-4012-b093-1307fd1211e9\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.919170 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8p65\" (UniqueName: \"kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65\") pod \"6303dec0-5e57-4012-b093-1307fd1211e9\" (UID: \"6303dec0-5e57-4012-b093-1307fd1211e9\") " Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.925874 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65" (OuterVolumeSpecName: "kube-api-access-p8p65") pod "6303dec0-5e57-4012-b093-1307fd1211e9" (UID: "6303dec0-5e57-4012-b093-1307fd1211e9"). InnerVolumeSpecName "kube-api-access-p8p65". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.951212 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config" (OuterVolumeSpecName: "config") pod "6303dec0-5e57-4012-b093-1307fd1211e9" (UID: "6303dec0-5e57-4012-b093-1307fd1211e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:52 crc kubenswrapper[4831]: I1124 08:32:52.965936 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6303dec0-5e57-4012-b093-1307fd1211e9" (UID: "6303dec0-5e57-4012-b093-1307fd1211e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.020924 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.020966 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6303dec0-5e57-4012-b093-1307fd1211e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.020981 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8p65\" (UniqueName: \"kubernetes.io/projected/6303dec0-5e57-4012-b093-1307fd1211e9-kube-api-access-p8p65\") on node \"crc\" DevicePath \"\"" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.599959 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerStarted","Data":"62dda8189a1f996b74c360ba067e784d3798db44d290011dfd82101d743904d8"} Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.601979 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nzfjn" event={"ID":"b76e6599-c355-4d06-80bd-cb66824a64a4","Type":"ContainerStarted","Data":"4f5617adfd564206c13fa008d2a5a73da35553619760045bd1b6c9fe926cc4a8"} Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.606787 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qqjkx" event={"ID":"d698dc95-8dfe-483d-82ac-21584ce29659","Type":"ContainerStarted","Data":"0236ca4597752c7d08649added42ad7ba113929de0b2b97eec87202ca72f1f0f"} Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.619346 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vm69f" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.619503 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vm69f" event={"ID":"6303dec0-5e57-4012-b093-1307fd1211e9","Type":"ContainerDied","Data":"d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6"} Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.619538 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d83ece19e9a8178f966d5e1196cab15d7d1925ea4b76322f43feab6433c86cb6" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.637567 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tv9wr" event={"ID":"b702363c-0c9b-4804-a478-01ea54210d2c","Type":"ContainerStarted","Data":"b38718cb79dbb610d00c953a9bdfd486986e181488e61a7e4e25567f74d50f36"} Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.659001 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nzfjn" podStartSLOduration=3.587632756 podStartE2EDuration="25.658983242s" podCreationTimestamp="2025-11-24 08:32:28 +0000 UTC" firstStartedPulling="2025-11-24 08:32:29.935611496 +0000 UTC m=+1023.810756649" lastFinishedPulling="2025-11-24 08:32:52.006961982 +0000 UTC m=+1045.882107135" observedRunningTime="2025-11-24 08:32:53.627142124 +0000 UTC m=+1047.502287277" watchObservedRunningTime="2025-11-24 08:32:53.658983242 +0000 UTC m=+1047.534128395" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.702680 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qqjkx" podStartSLOduration=3.461842783 podStartE2EDuration="25.702657768s" podCreationTimestamp="2025-11-24 08:32:28 +0000 UTC" firstStartedPulling="2025-11-24 08:32:29.767269669 +0000 UTC m=+1023.642414822" lastFinishedPulling="2025-11-24 08:32:52.008084654 +0000 UTC m=+1045.883229807" observedRunningTime="2025-11-24 08:32:53.652877728 +0000 UTC m=+1047.528022881" watchObservedRunningTime="2025-11-24 08:32:53.702657768 +0000 UTC m=+1047.577802931" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.712697 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tv9wr" podStartSLOduration=16.712669653 podStartE2EDuration="16.712669653s" podCreationTimestamp="2025-11-24 08:32:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:53.682095511 +0000 UTC m=+1047.557240674" watchObservedRunningTime="2025-11-24 08:32:53.712669653 +0000 UTC m=+1047.587814806" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.841966 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:32:53 crc kubenswrapper[4831]: E1124 08:32:53.843451 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="dnsmasq-dns" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.843472 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="dnsmasq-dns" Nov 24 08:32:53 crc kubenswrapper[4831]: E1124 08:32:53.843488 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6303dec0-5e57-4012-b093-1307fd1211e9" containerName="neutron-db-sync" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.843494 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6303dec0-5e57-4012-b093-1307fd1211e9" containerName="neutron-db-sync" Nov 24 08:32:53 crc kubenswrapper[4831]: E1124 08:32:53.843524 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="init" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.843530 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="init" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.843682 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6303dec0-5e57-4012-b093-1307fd1211e9" containerName="neutron-db-sync" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.843715 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="7053a17c-f398-4bb0-bc6d-30718f4e4596" containerName="dnsmasq-dns" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.844753 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.880440 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.968591 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.968642 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.968690 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.968803 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmzr4\" (UniqueName: \"kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.968836 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.986532 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:32:53 crc kubenswrapper[4831]: I1124 08:32:53.989155 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.000115 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.000355 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xzmx2" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.000498 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.000681 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.009542 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073106 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073156 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073175 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmzr4\" (UniqueName: \"kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073253 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073288 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fds5k\" (UniqueName: \"kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073331 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073357 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073376 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073400 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.073447 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.074009 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.074962 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.078656 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.078660 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.131494 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmzr4\" (UniqueName: \"kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4\") pod \"dnsmasq-dns-5f66db59b9-b4rph\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.167900 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.175645 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.175713 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.175790 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.175841 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fds5k\" (UniqueName: \"kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.175890 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.198107 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fds5k\" (UniqueName: \"kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.203214 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.204918 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.211187 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.213796 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs\") pod \"neutron-86987b4f86-lqghd\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:54 crc kubenswrapper[4831]: I1124 08:32:54.319727 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:32:55 crc kubenswrapper[4831]: I1124 08:32:54.733281 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:32:55 crc kubenswrapper[4831]: I1124 08:32:55.679727 4831 generic.go:334] "Generic (PLEG): container finished" podID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerID="b79b44d0793c70957c8fb43ac6f41e3cfda9017c1a458640b9ea9233b462190c" exitCode=0 Nov 24 08:32:55 crc kubenswrapper[4831]: I1124 08:32:55.679795 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" event={"ID":"c9a03652-b31c-46ae-b1c4-2539f51e44d2","Type":"ContainerDied","Data":"b79b44d0793c70957c8fb43ac6f41e3cfda9017c1a458640b9ea9233b462190c"} Nov 24 08:32:55 crc kubenswrapper[4831]: I1124 08:32:55.679819 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" event={"ID":"c9a03652-b31c-46ae-b1c4-2539f51e44d2","Type":"ContainerStarted","Data":"26609b806307a2402e9d3677fc34e4c36b2f87fa54b313a1041c3d10c2a3aec9"} Nov 24 08:32:56 crc kubenswrapper[4831]: I1124 08:32:56.242143 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:32:56 crc kubenswrapper[4831]: I1124 08:32:56.710255 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" event={"ID":"c9a03652-b31c-46ae-b1c4-2539f51e44d2","Type":"ContainerStarted","Data":"4c93f6fa433e3fe553b087bc43dbc31c1fb5493873b9e7c61a6fecf1730940e3"} Nov 24 08:32:56 crc kubenswrapper[4831]: I1124 08:32:56.710526 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:32:56 crc kubenswrapper[4831]: I1124 08:32:56.756370 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" podStartSLOduration=3.7563509870000003 podStartE2EDuration="3.756350987s" podCreationTimestamp="2025-11-24 08:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:32:56.748490562 +0000 UTC m=+1050.623635715" watchObservedRunningTime="2025-11-24 08:32:56.756350987 +0000 UTC m=+1050.631496130" Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.720127 4831 generic.go:334] "Generic (PLEG): container finished" podID="b76e6599-c355-4d06-80bd-cb66824a64a4" containerID="4f5617adfd564206c13fa008d2a5a73da35553619760045bd1b6c9fe926cc4a8" exitCode=0 Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.721132 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nzfjn" event={"ID":"b76e6599-c355-4d06-80bd-cb66824a64a4","Type":"ContainerDied","Data":"4f5617adfd564206c13fa008d2a5a73da35553619760045bd1b6c9fe926cc4a8"} Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.904478 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79dd578589-cbqvb"] Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.914879 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.920979 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.922473 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79dd578589-cbqvb"] Nov 24 08:32:57 crc kubenswrapper[4831]: I1124 08:32:57.922531 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.056978 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-httpd-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057031 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-internal-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057063 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057127 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-combined-ca-bundle\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057150 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-public-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057247 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4mp2\" (UniqueName: \"kubernetes.io/projected/a511397f-1a08-4819-8f25-4435c4e5fe6a-kube-api-access-t4mp2\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.057269 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-ovndb-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158661 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-httpd-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158719 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-internal-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158753 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158810 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-combined-ca-bundle\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158838 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-public-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158911 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4mp2\" (UniqueName: \"kubernetes.io/projected/a511397f-1a08-4819-8f25-4435c4e5fe6a-kube-api-access-t4mp2\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.158939 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-ovndb-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.167638 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-internal-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.167934 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-ovndb-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.168020 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-combined-ca-bundle\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.172351 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-public-tls-certs\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.175442 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-httpd-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.177288 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4mp2\" (UniqueName: \"kubernetes.io/projected/a511397f-1a08-4819-8f25-4435c4e5fe6a-kube-api-access-t4mp2\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.185183 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a511397f-1a08-4819-8f25-4435c4e5fe6a-config\") pod \"neutron-79dd578589-cbqvb\" (UID: \"a511397f-1a08-4819-8f25-4435c4e5fe6a\") " pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.254227 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:32:58 crc kubenswrapper[4831]: W1124 08:32:58.474983 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9cde5cb_8b3b_43f6_a7f1_49ee4f25865e.slice/crio-61ed820e2070f79c2ce479c4bcea659a7152f8475c86da3df899e6c05865a7de WatchSource:0}: Error finding container 61ed820e2070f79c2ce479c4bcea659a7152f8475c86da3df899e6c05865a7de: Status 404 returned error can't find the container with id 61ed820e2070f79c2ce479c4bcea659a7152f8475c86da3df899e6c05865a7de Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.737639 4831 generic.go:334] "Generic (PLEG): container finished" podID="d698dc95-8dfe-483d-82ac-21584ce29659" containerID="0236ca4597752c7d08649added42ad7ba113929de0b2b97eec87202ca72f1f0f" exitCode=0 Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.737770 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qqjkx" event={"ID":"d698dc95-8dfe-483d-82ac-21584ce29659","Type":"ContainerDied","Data":"0236ca4597752c7d08649added42ad7ba113929de0b2b97eec87202ca72f1f0f"} Nov 24 08:32:58 crc kubenswrapper[4831]: I1124 08:32:58.739040 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerStarted","Data":"61ed820e2070f79c2ce479c4bcea659a7152f8475c86da3df899e6c05865a7de"} Nov 24 08:32:59 crc kubenswrapper[4831]: I1124 08:32:59.751953 4831 generic.go:334] "Generic (PLEG): container finished" podID="b702363c-0c9b-4804-a478-01ea54210d2c" containerID="b38718cb79dbb610d00c953a9bdfd486986e181488e61a7e4e25567f74d50f36" exitCode=0 Nov 24 08:32:59 crc kubenswrapper[4831]: I1124 08:32:59.752533 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tv9wr" event={"ID":"b702363c-0c9b-4804-a478-01ea54210d2c","Type":"ContainerDied","Data":"b38718cb79dbb610d00c953a9bdfd486986e181488e61a7e4e25567f74d50f36"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.490973 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.494081 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nzfjn" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.502859 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.636689 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637031 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637062 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle\") pod \"d698dc95-8dfe-483d-82ac-21584ce29659\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637114 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle\") pod \"b76e6599-c355-4d06-80bd-cb66824a64a4\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637146 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637224 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data\") pod \"d698dc95-8dfe-483d-82ac-21584ce29659\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637261 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs\") pod \"b76e6599-c355-4d06-80bd-cb66824a64a4\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637340 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52h7t\" (UniqueName: \"kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637366 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts\") pod \"b76e6599-c355-4d06-80bd-cb66824a64a4\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637403 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data\") pod \"b76e6599-c355-4d06-80bd-cb66824a64a4\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637432 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t68ps\" (UniqueName: \"kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps\") pod \"b76e6599-c355-4d06-80bd-cb66824a64a4\" (UID: \"b76e6599-c355-4d06-80bd-cb66824a64a4\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637475 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9s2s\" (UniqueName: \"kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s\") pod \"d698dc95-8dfe-483d-82ac-21584ce29659\" (UID: \"d698dc95-8dfe-483d-82ac-21584ce29659\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637497 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.637562 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys\") pod \"b702363c-0c9b-4804-a478-01ea54210d2c\" (UID: \"b702363c-0c9b-4804-a478-01ea54210d2c\") " Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.638709 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs" (OuterVolumeSpecName: "logs") pod "b76e6599-c355-4d06-80bd-cb66824a64a4" (UID: "b76e6599-c355-4d06-80bd-cb66824a64a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.640919 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.650256 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s" (OuterVolumeSpecName: "kube-api-access-p9s2s") pod "d698dc95-8dfe-483d-82ac-21584ce29659" (UID: "d698dc95-8dfe-483d-82ac-21584ce29659"). InnerVolumeSpecName "kube-api-access-p9s2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.651846 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps" (OuterVolumeSpecName: "kube-api-access-t68ps") pod "b76e6599-c355-4d06-80bd-cb66824a64a4" (UID: "b76e6599-c355-4d06-80bd-cb66824a64a4"). InnerVolumeSpecName "kube-api-access-t68ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.655696 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t" (OuterVolumeSpecName: "kube-api-access-52h7t") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "kube-api-access-52h7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.667534 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.667608 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts" (OuterVolumeSpecName: "scripts") pod "b76e6599-c355-4d06-80bd-cb66824a64a4" (UID: "b76e6599-c355-4d06-80bd-cb66824a64a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.682546 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d698dc95-8dfe-483d-82ac-21584ce29659" (UID: "d698dc95-8dfe-483d-82ac-21584ce29659"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.689018 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts" (OuterVolumeSpecName: "scripts") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.693495 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data" (OuterVolumeSpecName: "config-data") pod "b76e6599-c355-4d06-80bd-cb66824a64a4" (UID: "b76e6599-c355-4d06-80bd-cb66824a64a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.693982 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data" (OuterVolumeSpecName: "config-data") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.695529 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b702363c-0c9b-4804-a478-01ea54210d2c" (UID: "b702363c-0c9b-4804-a478-01ea54210d2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.708966 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b76e6599-c355-4d06-80bd-cb66824a64a4" (UID: "b76e6599-c355-4d06-80bd-cb66824a64a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.709333 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d698dc95-8dfe-483d-82ac-21584ce29659" (UID: "d698dc95-8dfe-483d-82ac-21584ce29659"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741374 4831 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741406 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b76e6599-c355-4d06-80bd-cb66824a64a4-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741419 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52h7t\" (UniqueName: \"kubernetes.io/projected/b702363c-0c9b-4804-a478-01ea54210d2c-kube-api-access-52h7t\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741436 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741446 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741454 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t68ps\" (UniqueName: \"kubernetes.io/projected/b76e6599-c355-4d06-80bd-cb66824a64a4-kube-api-access-t68ps\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741463 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9s2s\" (UniqueName: \"kubernetes.io/projected/d698dc95-8dfe-483d-82ac-21584ce29659-kube-api-access-p9s2s\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741473 4831 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741482 4831 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741490 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741498 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d698dc95-8dfe-483d-82ac-21584ce29659-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741507 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741515 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b76e6599-c355-4d06-80bd-cb66824a64a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.741523 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b702363c-0c9b-4804-a478-01ea54210d2c-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.773141 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerStarted","Data":"d742976952a73d3bfce96a5d64de0ef9eb4482ea07af3ef0a0ca6fe52f16931f"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.773202 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerStarted","Data":"d529b16349b5d6a3383d438f6bfebb9abb75bbe26ecf56ffeff69709f5810db5"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.773220 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.775520 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerStarted","Data":"373921e7da5f558ae1a83d765f0c0b5b4325185f57479d28b7a946dfe342a0f9"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.777247 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nzfjn" event={"ID":"b76e6599-c355-4d06-80bd-cb66824a64a4","Type":"ContainerDied","Data":"bc667bd4399949ecea663531d2497f59bcf1e5eebec622e32ae63b1456d0e78c"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.777275 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc667bd4399949ecea663531d2497f59bcf1e5eebec622e32ae63b1456d0e78c" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.777349 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nzfjn" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.781749 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qqjkx" event={"ID":"d698dc95-8dfe-483d-82ac-21584ce29659","Type":"ContainerDied","Data":"7fbf1620f4cb40ff91142f2191ed1702103fdd49c5f5ad8054f561afe1f2602c"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.781791 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fbf1620f4cb40ff91142f2191ed1702103fdd49c5f5ad8054f561afe1f2602c" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.781855 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qqjkx" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.788779 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tv9wr" event={"ID":"b702363c-0c9b-4804-a478-01ea54210d2c","Type":"ContainerDied","Data":"58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe"} Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.788812 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58e550706bfb9ef8f04c2ac066155fdc5489e719323ec61fd26548b63726bbfe" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.788867 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tv9wr" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.804336 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-86987b4f86-lqghd" podStartSLOduration=8.804298344 podStartE2EDuration="8.804298344s" podCreationTimestamp="2025-11-24 08:32:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:01.795613827 +0000 UTC m=+1055.670758980" watchObservedRunningTime="2025-11-24 08:33:01.804298344 +0000 UTC m=+1055.679443497" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.952191 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79dd578589-cbqvb"] Nov 24 08:33:01 crc kubenswrapper[4831]: W1124 08:33:01.952875 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda511397f_1a08_4819_8f25_4435c4e5fe6a.slice/crio-6ff73641d59c0fd6fda29c2bcb380e77a8f1099d31ba4be58c05e8efd7f35a42 WatchSource:0}: Error finding container 6ff73641d59c0fd6fda29c2bcb380e77a8f1099d31ba4be58c05e8efd7f35a42: Status 404 returned error can't find the container with id 6ff73641d59c0fd6fda29c2bcb380e77a8f1099d31ba4be58c05e8efd7f35a42 Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.976728 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-77b5948bf-z797k"] Nov 24 08:33:01 crc kubenswrapper[4831]: E1124 08:33:01.977197 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b76e6599-c355-4d06-80bd-cb66824a64a4" containerName="placement-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.977283 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b76e6599-c355-4d06-80bd-cb66824a64a4" containerName="placement-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: E1124 08:33:01.977408 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d698dc95-8dfe-483d-82ac-21584ce29659" containerName="barbican-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.977569 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d698dc95-8dfe-483d-82ac-21584ce29659" containerName="barbican-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: E1124 08:33:01.977657 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b702363c-0c9b-4804-a478-01ea54210d2c" containerName="keystone-bootstrap" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.977718 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b702363c-0c9b-4804-a478-01ea54210d2c" containerName="keystone-bootstrap" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.977987 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b702363c-0c9b-4804-a478-01ea54210d2c" containerName="keystone-bootstrap" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.978063 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b76e6599-c355-4d06-80bd-cb66824a64a4" containerName="placement-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.978143 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d698dc95-8dfe-483d-82ac-21584ce29659" containerName="barbican-db-sync" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.978914 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.992151 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.992416 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.993099 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.993530 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-m2wp6" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.995966 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 24 08:33:01 crc kubenswrapper[4831]: I1124 08:33:01.996256 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.000767 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77b5948bf-z797k"] Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.047680 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-credential-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.047738 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-config-data\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.047760 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-public-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.048503 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-internal-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.048576 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-scripts\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.048647 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-combined-ca-bundle\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.048762 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-fernet-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.048815 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlspk\" (UniqueName: \"kubernetes.io/projected/84b88081-1ab2-4e86-a37d-3b4687aeb91e-kube-api-access-qlspk\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.150736 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-internal-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.150814 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-scripts\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.150855 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-combined-ca-bundle\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.150928 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-fernet-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.150971 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlspk\" (UniqueName: \"kubernetes.io/projected/84b88081-1ab2-4e86-a37d-3b4687aeb91e-kube-api-access-qlspk\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.151006 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-credential-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.151048 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-config-data\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.151073 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-public-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.154746 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-scripts\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.155680 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-credential-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.156099 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-combined-ca-bundle\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.160020 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-internal-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.160254 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-fernet-keys\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.160379 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-config-data\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.166835 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/84b88081-1ab2-4e86-a37d-3b4687aeb91e-public-tls-certs\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.178663 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlspk\" (UniqueName: \"kubernetes.io/projected/84b88081-1ab2-4e86-a37d-3b4687aeb91e-kube-api-access-qlspk\") pod \"keystone-77b5948bf-z797k\" (UID: \"84b88081-1ab2-4e86-a37d-3b4687aeb91e\") " pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.316406 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.703983 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-64c5f59669-7kdzn"] Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.728949 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64c5f59669-7kdzn"] Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.729233 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.737821 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fjjgc" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.737821 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.737988 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.838432 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79dd578589-cbqvb" event={"ID":"a511397f-1a08-4819-8f25-4435c4e5fe6a","Type":"ContainerStarted","Data":"741f5567644dba1370f516b3ca6c4aba2beb4f54e40f857ba45cf5fcf9e99f0c"} Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.838699 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79dd578589-cbqvb" event={"ID":"a511397f-1a08-4819-8f25-4435c4e5fe6a","Type":"ContainerStarted","Data":"da3f64d811299807290e02e2139c318a31e0130d0f6ee6d32b75647e7b4908b9"} Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.838772 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79dd578589-cbqvb" event={"ID":"a511397f-1a08-4819-8f25-4435c4e5fe6a","Type":"ContainerStarted","Data":"6ff73641d59c0fd6fda29c2bcb380e77a8f1099d31ba4be58c05e8efd7f35a42"} Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.839953 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.860464 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-combined-ca-bundle\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.861030 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.861100 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1087935c-a660-42ae-bc8e-dc8c91bbb414-logs\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.861179 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data-custom\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.861271 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9klt\" (UniqueName: \"kubernetes.io/projected/1087935c-a660-42ae-bc8e-dc8c91bbb414-kube-api-access-h9klt\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.862588 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7bcf554d66-64cr7"] Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.863937 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.894119 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/0.log" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.894361 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.899463 4831 generic.go:334] "Generic (PLEG): container finished" podID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerID="d742976952a73d3bfce96a5d64de0ef9eb4482ea07af3ef0a0ca6fe52f16931f" exitCode=1 Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.900345 4831 scope.go:117] "RemoveContainer" containerID="d742976952a73d3bfce96a5d64de0ef9eb4482ea07af3ef0a0ca6fe52f16931f" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.915096 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79dd578589-cbqvb" podStartSLOduration=5.915080709 podStartE2EDuration="5.915080709s" podCreationTimestamp="2025-11-24 08:32:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:02.880569455 +0000 UTC m=+1056.755714618" watchObservedRunningTime="2025-11-24 08:33:02.915080709 +0000 UTC m=+1056.790225862" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.915853 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6f67674df6-sk4ws"] Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.916984 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerDied","Data":"d742976952a73d3bfce96a5d64de0ef9eb4482ea07af3ef0a0ca6fe52f16931f"} Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.917069 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.922682 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.923094 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.923332 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.931831 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.931949 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-vjpmw" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975583 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-combined-ca-bundle\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975665 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8a5828-9ea3-402e-812d-5dbbbe749087-logs\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975696 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975731 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1087935c-a660-42ae-bc8e-dc8c91bbb414-logs\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975764 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmj9c\" (UniqueName: \"kubernetes.io/projected/6b8a5828-9ea3-402e-812d-5dbbbe749087-kube-api-access-qmj9c\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975817 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975838 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data-custom\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975883 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9klt\" (UniqueName: \"kubernetes.io/projected/1087935c-a660-42ae-bc8e-dc8c91bbb414-kube-api-access-h9klt\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975938 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data-custom\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.975988 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-combined-ca-bundle\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:02 crc kubenswrapper[4831]: I1124 08:33:02.998555 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.006933 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="dnsmasq-dns" containerID="cri-o://4c93f6fa433e3fe553b087bc43dbc31c1fb5493873b9e7c61a6fecf1730940e3" gracePeriod=10 Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.010507 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.014822 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1087935c-a660-42ae-bc8e-dc8c91bbb414-logs\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.040917 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data-custom\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.041963 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-config-data\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.065165 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1087935c-a660-42ae-bc8e-dc8c91bbb414-combined-ca-bundle\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.095227 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bcf554d66-64cr7"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.097512 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.097748 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-combined-ca-bundle\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.097834 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data-custom\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.097920 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p579l\" (UniqueName: \"kubernetes.io/projected/610d7193-b921-486e-8d2c-aa897a589c30-kube-api-access-p579l\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.097997 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-config-data\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098100 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-internal-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098208 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-combined-ca-bundle\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098286 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610d7193-b921-486e-8d2c-aa897a589c30-logs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-scripts\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098441 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-public-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098529 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8a5828-9ea3-402e-812d-5dbbbe749087-logs\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.098608 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmj9c\" (UniqueName: \"kubernetes.io/projected/6b8a5828-9ea3-402e-812d-5dbbbe749087-kube-api-access-qmj9c\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.109971 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8a5828-9ea3-402e-812d-5dbbbe749087-logs\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.127236 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data-custom\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.134445 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-combined-ca-bundle\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.138334 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9klt\" (UniqueName: \"kubernetes.io/projected/1087935c-a660-42ae-bc8e-dc8c91bbb414-kube-api-access-h9klt\") pod \"barbican-worker-64c5f59669-7kdzn\" (UID: \"1087935c-a660-42ae-bc8e-dc8c91bbb414\") " pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.154490 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8a5828-9ea3-402e-812d-5dbbbe749087-config-data\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.173513 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmj9c\" (UniqueName: \"kubernetes.io/projected/6b8a5828-9ea3-402e-812d-5dbbbe749087-kube-api-access-qmj9c\") pod \"barbican-keystone-listener-7bcf554d66-64cr7\" (UID: \"6b8a5828-9ea3-402e-812d-5dbbbe749087\") " pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207480 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610d7193-b921-486e-8d2c-aa897a589c30-logs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207533 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-scripts\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207557 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-public-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207619 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-combined-ca-bundle\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207653 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p579l\" (UniqueName: \"kubernetes.io/projected/610d7193-b921-486e-8d2c-aa897a589c30-kube-api-access-p579l\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207669 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-config-data\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.207753 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-internal-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.214012 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-internal-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.214272 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610d7193-b921-486e-8d2c-aa897a589c30-logs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.214713 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.217031 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f67674df6-sk4ws"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.223896 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77b5948bf-z797k"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.241032 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-config-data\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.258036 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-combined-ca-bundle\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.300334 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.302209 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.303827 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-public-tls-certs\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.304062 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/610d7193-b921-486e-8d2c-aa897a589c30-scripts\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.335483 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.348457 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p579l\" (UniqueName: \"kubernetes.io/projected/610d7193-b921-486e-8d2c-aa897a589c30-kube-api-access-p579l\") pod \"placement-6f67674df6-sk4ws\" (UID: \"610d7193-b921-486e-8d2c-aa897a589c30\") " pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.355176 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-64c5f59669-7kdzn" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.413980 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9swmv\" (UniqueName: \"kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.414043 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.414069 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.414086 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.414107 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.516565 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.516626 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.516651 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.516764 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9swmv\" (UniqueName: \"kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.516811 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.517813 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.518186 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.518379 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.518907 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.528834 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.530728 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.533622 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.569539 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.576375 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9swmv\" (UniqueName: \"kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv\") pod \"dnsmasq-dns-869f779d85-cmjxr\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.620229 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.620503 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2rrp\" (UniqueName: \"kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.620680 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.645201 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.646000 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.659384 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.670514 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.782252 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.783058 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.783177 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.783523 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.783592 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2rrp\" (UniqueName: \"kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.784598 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.824581 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2rrp\" (UniqueName: \"kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.836736 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.842481 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:03 crc kubenswrapper[4831]: I1124 08:33:03.843095 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle\") pod \"barbican-api-5f7b65778-ck4pk\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.024797 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.052453 4831 generic.go:334] "Generic (PLEG): container finished" podID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerID="4c93f6fa433e3fe553b087bc43dbc31c1fb5493873b9e7c61a6fecf1730940e3" exitCode=0 Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.052567 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" event={"ID":"c9a03652-b31c-46ae-b1c4-2539f51e44d2","Type":"ContainerDied","Data":"4c93f6fa433e3fe553b087bc43dbc31c1fb5493873b9e7c61a6fecf1730940e3"} Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.076297 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77b5948bf-z797k" event={"ID":"84b88081-1ab2-4e86-a37d-3b4687aeb91e","Type":"ContainerStarted","Data":"56bedd915c36e4818176998e7a229cbad1a16fa66170d490d0e5e7b76e351c25"} Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.206703 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.361692 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb\") pod \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.362092 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb\") pod \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.362280 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc\") pod \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.362365 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmzr4\" (UniqueName: \"kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4\") pod \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.362401 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config\") pod \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\" (UID: \"c9a03652-b31c-46ae-b1c4-2539f51e44d2\") " Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.426004 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-64c5f59669-7kdzn"] Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.459711 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4" (OuterVolumeSpecName: "kube-api-access-vmzr4") pod "c9a03652-b31c-46ae-b1c4-2539f51e44d2" (UID: "c9a03652-b31c-46ae-b1c4-2539f51e44d2"). InnerVolumeSpecName "kube-api-access-vmzr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.468812 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmzr4\" (UniqueName: \"kubernetes.io/projected/c9a03652-b31c-46ae-b1c4-2539f51e44d2-kube-api-access-vmzr4\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.705831 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9a03652-b31c-46ae-b1c4-2539f51e44d2" (UID: "c9a03652-b31c-46ae-b1c4-2539f51e44d2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.761950 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bcf554d66-64cr7"] Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.783365 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9a03652-b31c-46ae-b1c4-2539f51e44d2" (UID: "c9a03652-b31c-46ae-b1c4-2539f51e44d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.785116 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.785138 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.804604 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6f67674df6-sk4ws"] Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.831911 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config" (OuterVolumeSpecName: "config") pod "c9a03652-b31c-46ae-b1c4-2539f51e44d2" (UID: "c9a03652-b31c-46ae-b1c4-2539f51e44d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.852155 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9a03652-b31c-46ae-b1c4-2539f51e44d2" (UID: "c9a03652-b31c-46ae-b1c4-2539f51e44d2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.886390 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.886421 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9a03652-b31c-46ae-b1c4-2539f51e44d2-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.917970 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:04 crc kubenswrapper[4831]: I1124 08:33:04.960138 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.096985 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" event={"ID":"6b8a5828-9ea3-402e-812d-5dbbbe749087","Type":"ContainerStarted","Data":"18984b75deed86f29f3818bfae2a94d8215c81f00008330f151228d4a0742088"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.104441 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerStarted","Data":"f369f6553f77d6afb94446412b62134da4985825bc18323168dec2f5fc230f7a"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.114751 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f67674df6-sk4ws" event={"ID":"610d7193-b921-486e-8d2c-aa897a589c30","Type":"ContainerStarted","Data":"22c257d49431abf8210c5167e4b52896d478aa8ccadee843af30b2849adc1f0e"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.122511 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" event={"ID":"c9a03652-b31c-46ae-b1c4-2539f51e44d2","Type":"ContainerDied","Data":"26609b806307a2402e9d3677fc34e4c36b2f87fa54b313a1041c3d10c2a3aec9"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.122561 4831 scope.go:117] "RemoveContainer" containerID="4c93f6fa433e3fe553b087bc43dbc31c1fb5493873b9e7c61a6fecf1730940e3" Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.122719 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.129761 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64c5f59669-7kdzn" event={"ID":"1087935c-a660-42ae-bc8e-dc8c91bbb414","Type":"ContainerStarted","Data":"5972f610132bc48ada8eb039970096f399fa16a09bfd27643446ff4937c8497a"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.150462 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77b5948bf-z797k" event={"ID":"84b88081-1ab2-4e86-a37d-3b4687aeb91e","Type":"ContainerStarted","Data":"41d6b88d303cfc5cf8b3f278391b975ddccebce9bcf64535acf7cf9f41795bbc"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.159785 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" event={"ID":"821e46e4-876a-4b97-b4be-70baecc6b883","Type":"ContainerStarted","Data":"4ecf8a4f87efb05d402a4d09b2ca17cdbc44c62ff79fbf6e76aa98302efe3b35"} Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.164677 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.177889 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-b4rph"] Nov 24 08:33:05 crc kubenswrapper[4831]: I1124 08:33:05.204754 4831 scope.go:117] "RemoveContainer" containerID="b79b44d0793c70957c8fb43ac6f41e3cfda9017c1a458640b9ea9233b462190c" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.226256 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerStarted","Data":"5446ac08ad4e1d5be18e412e6b44963d742b99d385de1b4b1d2ab3138bcda82d"} Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.284607 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d4cd7dbc6-zgxq2"] Nov 24 08:33:06 crc kubenswrapper[4831]: E1124 08:33:06.292701 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="init" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.292736 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="init" Nov 24 08:33:06 crc kubenswrapper[4831]: E1124 08:33:06.292746 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="dnsmasq-dns" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.292752 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="dnsmasq-dns" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.292936 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="dnsmasq-dns" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.293787 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f67674df6-sk4ws" event={"ID":"610d7193-b921-486e-8d2c-aa897a589c30","Type":"ContainerStarted","Data":"53b423a948f647396b2078e5a1813f61ecd61293b2001622da856ec562711e43"} Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.293882 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.309965 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.309965 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.311130 4831 generic.go:334] "Generic (PLEG): container finished" podID="821e46e4-876a-4b97-b4be-70baecc6b883" containerID="6d7e84c26d3fcf180c0c88849374e1fd9aad2d2a6a76c0aa84cedbd744e10322" exitCode=0 Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.311189 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" event={"ID":"821e46e4-876a-4b97-b4be-70baecc6b883","Type":"ContainerDied","Data":"6d7e84c26d3fcf180c0c88849374e1fd9aad2d2a6a76c0aa84cedbd744e10322"} Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.320710 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/0.log" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327409 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-combined-ca-bundle\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327540 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data-custom\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327563 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327593 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-internal-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327625 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mx8j\" (UniqueName: \"kubernetes.io/projected/52ce26fc-b850-4fb1-9539-e9599f080475-kube-api-access-5mx8j\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327677 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-public-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.327710 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52ce26fc-b850-4fb1-9539-e9599f080475-logs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.335500 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerStarted","Data":"a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae"} Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.335546 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.336435 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.346578 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d4cd7dbc6-zgxq2"] Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431475 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-combined-ca-bundle\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431567 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data-custom\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431586 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431615 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-internal-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431644 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mx8j\" (UniqueName: \"kubernetes.io/projected/52ce26fc-b850-4fb1-9539-e9599f080475-kube-api-access-5mx8j\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431685 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-public-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.431711 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52ce26fc-b850-4fb1-9539-e9599f080475-logs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.432816 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/52ce26fc-b850-4fb1-9539-e9599f080475-logs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.458732 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.459627 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-internal-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.460377 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-config-data-custom\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.464396 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-combined-ca-bundle\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.465093 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/52ce26fc-b850-4fb1-9539-e9599f080475-public-tls-certs\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.493223 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mx8j\" (UniqueName: \"kubernetes.io/projected/52ce26fc-b850-4fb1-9539-e9599f080475-kube-api-access-5mx8j\") pod \"barbican-api-6d4cd7dbc6-zgxq2\" (UID: \"52ce26fc-b850-4fb1-9539-e9599f080475\") " pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.508977 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.517528 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-77b5948bf-z797k" podStartSLOduration=5.517508846 podStartE2EDuration="5.517508846s" podCreationTimestamp="2025-11-24 08:33:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:06.510617719 +0000 UTC m=+1060.385762882" watchObservedRunningTime="2025-11-24 08:33:06.517508846 +0000 UTC m=+1060.392653999" Nov 24 08:33:06 crc kubenswrapper[4831]: I1124 08:33:06.917586 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" path="/var/lib/kubelet/pods/c9a03652-b31c-46ae-b1c4-2539f51e44d2/volumes" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.164230 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d4cd7dbc6-zgxq2"] Nov 24 08:33:07 crc kubenswrapper[4831]: W1124 08:33:07.201264 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52ce26fc_b850_4fb1_9539_e9599f080475.slice/crio-e7a5f82fabddeb8cc51a3bbf0ee85056ec6487c5c84a7e519a81dee2e94f862e WatchSource:0}: Error finding container e7a5f82fabddeb8cc51a3bbf0ee85056ec6487c5c84a7e519a81dee2e94f862e: Status 404 returned error can't find the container with id e7a5f82fabddeb8cc51a3bbf0ee85056ec6487c5c84a7e519a81dee2e94f862e Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.376942 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" event={"ID":"821e46e4-876a-4b97-b4be-70baecc6b883","Type":"ContainerStarted","Data":"02892049a4e10c5a353df4ae463fac74d7a38b64ab923c6b87dbce7be0772d85"} Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.377306 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.391765 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/1.log" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.392161 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/0.log" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.394020 4831 generic.go:334] "Generic (PLEG): container finished" podID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerID="a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae" exitCode=1 Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.394079 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerDied","Data":"a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae"} Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.394113 4831 scope.go:117] "RemoveContainer" containerID="d742976952a73d3bfce96a5d64de0ef9eb4482ea07af3ef0a0ca6fe52f16931f" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.394813 4831 scope.go:117] "RemoveContainer" containerID="a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae" Nov 24 08:33:07 crc kubenswrapper[4831]: E1124 08:33:07.394999 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 10s restarting failed container=neutron-httpd pod=neutron-86987b4f86-lqghd_openstack(d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e)\"" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.404061 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" podStartSLOduration=5.404043777 podStartE2EDuration="5.404043777s" podCreationTimestamp="2025-11-24 08:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:07.39681086 +0000 UTC m=+1061.271956023" watchObservedRunningTime="2025-11-24 08:33:07.404043777 +0000 UTC m=+1061.279188930" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.415687 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" event={"ID":"52ce26fc-b850-4fb1-9539-e9599f080475","Type":"ContainerStarted","Data":"e7a5f82fabddeb8cc51a3bbf0ee85056ec6487c5c84a7e519a81dee2e94f862e"} Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.421849 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerStarted","Data":"28a8488b69d05767971f9633c1d7b2c5493726941f0b04d6f975d59f2c8ce115"} Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.423186 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.423360 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.424789 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6f67674df6-sk4ws" event={"ID":"610d7193-b921-486e-8d2c-aa897a589c30","Type":"ContainerStarted","Data":"511763ef72ec42b7fd77339fd653e93b73ecbd251c4c1733073b914e8c1c2a89"} Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.424842 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.424925 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.482895 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6f67674df6-sk4ws" podStartSLOduration=5.482880715 podStartE2EDuration="5.482880715s" podCreationTimestamp="2025-11-24 08:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:07.482128423 +0000 UTC m=+1061.357273576" watchObservedRunningTime="2025-11-24 08:33:07.482880715 +0000 UTC m=+1061.358025868" Nov 24 08:33:07 crc kubenswrapper[4831]: I1124 08:33:07.513192 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f7b65778-ck4pk" podStartSLOduration=4.513173519 podStartE2EDuration="4.513173519s" podCreationTimestamp="2025-11-24 08:33:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:07.509972657 +0000 UTC m=+1061.385117820" watchObservedRunningTime="2025-11-24 08:33:07.513173519 +0000 UTC m=+1061.388318662" Nov 24 08:33:08 crc kubenswrapper[4831]: I1124 08:33:08.444176 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" event={"ID":"52ce26fc-b850-4fb1-9539-e9599f080475","Type":"ContainerStarted","Data":"d7af9c75251c0b365cd3dc6467d778c7c33d6b5ca1dc336ad9af274e9caaf00c"} Nov 24 08:33:08 crc kubenswrapper[4831]: I1124 08:33:08.444836 4831 scope.go:117] "RemoveContainer" containerID="a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae" Nov 24 08:33:08 crc kubenswrapper[4831]: E1124 08:33:08.445069 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 10s restarting failed container=neutron-httpd pod=neutron-86987b4f86-lqghd_openstack(d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e)\"" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" Nov 24 08:33:09 crc kubenswrapper[4831]: I1124 08:33:09.171425 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f66db59b9-b4rph" podUID="c9a03652-b31c-46ae-b1c4-2539f51e44d2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Nov 24 08:33:09 crc kubenswrapper[4831]: I1124 08:33:09.456079 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/1.log" Nov 24 08:33:13 crc kubenswrapper[4831]: I1124 08:33:13.661303 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:13 crc kubenswrapper[4831]: I1124 08:33:13.733392 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:33:13 crc kubenswrapper[4831]: I1124 08:33:13.735800 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="dnsmasq-dns" containerID="cri-o://2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4" gracePeriod=10 Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.010214 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Nov 24 08:33:14 crc kubenswrapper[4831]: E1124 08:33:14.430213 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59446fe2_a549_4797_8fbd_370904369dd4.slice/crio-2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59446fe2_a549_4797_8fbd_370904369dd4.slice/crio-conmon-2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.595225 4831 generic.go:334] "Generic (PLEG): container finished" podID="59446fe2-a549-4797-8fbd-370904369dd4" containerID="2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4" exitCode=0 Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.595773 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" event={"ID":"59446fe2-a549-4797-8fbd-370904369dd4","Type":"ContainerDied","Data":"2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4"} Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.622147 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64c5f59669-7kdzn" event={"ID":"1087935c-a660-42ae-bc8e-dc8c91bbb414","Type":"ContainerStarted","Data":"fab5363116cac803a1f4c6b6e1c2e1055b7ecbadd61e8152fb5d549e757d4284"} Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.668093 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.668745 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.693611 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" podUID="52ce26fc-b850-4fb1-9539-e9599f080475" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.150:9311/healthcheck\": dial tcp 10.217.0.150:9311: connect: connection refused" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.699231 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" event={"ID":"6b8a5828-9ea3-402e-812d-5dbbbe749087","Type":"ContainerStarted","Data":"7ae84ec527a320d42da22512342b499c0d9758f72fa5a3a3f655a332e9f322bf"} Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.699500 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.722820 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" podStartSLOduration=8.722802799 podStartE2EDuration="8.722802799s" podCreationTimestamp="2025-11-24 08:33:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:14.718595899 +0000 UTC m=+1068.593741052" watchObservedRunningTime="2025-11-24 08:33:14.722802799 +0000 UTC m=+1068.597947952" Nov 24 08:33:14 crc kubenswrapper[4831]: E1124 08:33:14.753537 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.883027 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb\") pod \"59446fe2-a549-4797-8fbd-370904369dd4\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.883506 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc\") pod \"59446fe2-a549-4797-8fbd-370904369dd4\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.883538 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config\") pod \"59446fe2-a549-4797-8fbd-370904369dd4\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.883596 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpz7m\" (UniqueName: \"kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m\") pod \"59446fe2-a549-4797-8fbd-370904369dd4\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.883671 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb\") pod \"59446fe2-a549-4797-8fbd-370904369dd4\" (UID: \"59446fe2-a549-4797-8fbd-370904369dd4\") " Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.895331 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m" (OuterVolumeSpecName: "kube-api-access-wpz7m") pod "59446fe2-a549-4797-8fbd-370904369dd4" (UID: "59446fe2-a549-4797-8fbd-370904369dd4"). InnerVolumeSpecName "kube-api-access-wpz7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:14 crc kubenswrapper[4831]: I1124 08:33:14.986149 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpz7m\" (UniqueName: \"kubernetes.io/projected/59446fe2-a549-4797-8fbd-370904369dd4-kube-api-access-wpz7m\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.003543 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59446fe2-a549-4797-8fbd-370904369dd4" (UID: "59446fe2-a549-4797-8fbd-370904369dd4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.009538 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59446fe2-a549-4797-8fbd-370904369dd4" (UID: "59446fe2-a549-4797-8fbd-370904369dd4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.025336 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config" (OuterVolumeSpecName: "config") pod "59446fe2-a549-4797-8fbd-370904369dd4" (UID: "59446fe2-a549-4797-8fbd-370904369dd4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.026193 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59446fe2-a549-4797-8fbd-370904369dd4" (UID: "59446fe2-a549-4797-8fbd-370904369dd4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.090051 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.090102 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.090118 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.090132 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59446fe2-a549-4797-8fbd-370904369dd4-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.708727 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerStarted","Data":"d71f8a55572711739f001d9af6310dd8a6baeb6d675851090dca851564493f9e"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.710236 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.709568 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="proxy-httpd" containerID="cri-o://d71f8a55572711739f001d9af6310dd8a6baeb6d675851090dca851564493f9e" gracePeriod=30 Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.709584 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="sg-core" containerID="cri-o://373921e7da5f558ae1a83d765f0c0b5b4325185f57479d28b7a946dfe342a0f9" gracePeriod=30 Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.709092 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="ceilometer-notification-agent" containerID="cri-o://62dda8189a1f996b74c360ba067e784d3798db44d290011dfd82101d743904d8" gracePeriod=30 Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.712933 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" event={"ID":"52ce26fc-b850-4fb1-9539-e9599f080475","Type":"ContainerStarted","Data":"30f1bef1b3c1a1d9d16e538c5c620534390237c4c3ad6a3a72db32140662d2bb"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.719050 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" event={"ID":"6b8a5828-9ea3-402e-812d-5dbbbe749087","Type":"ContainerStarted","Data":"b5c61454db8bf43bf127a54006fb27552c041bec2c2df7e2fba259c5e694d461"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.720852 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" event={"ID":"59446fe2-a549-4797-8fbd-370904369dd4","Type":"ContainerDied","Data":"26cdb90bc8af95515f1216a382201aac22172459d2c1c86f44b3b20f5e6cbe89"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.720894 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-jstgx" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.720916 4831 scope.go:117] "RemoveContainer" containerID="2cf2fada0b5f77f08437e586740ef736ff7c93bcda2600ccbc25e013b0be2dd4" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.722720 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tqhj9" event={"ID":"46a54561-e390-48c6-a27d-e516a013ebb3","Type":"ContainerStarted","Data":"2c1339dbdb125375189623fed665b0cf1f79e5068870958ef2c0c843ac1f510c"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.726477 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-64c5f59669-7kdzn" event={"ID":"1087935c-a660-42ae-bc8e-dc8c91bbb414","Type":"ContainerStarted","Data":"cef40392d8e5c5e79febca0b92ca1577594070343db140bd3323a8f237bcc4cf"} Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.753127 4831 scope.go:117] "RemoveContainer" containerID="859aab855c9969fbbbd1cd147ed965f97eb96360b69d06469885cf4f0f695f1c" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.806365 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7bcf554d66-64cr7" podStartSLOduration=4.993612993 podStartE2EDuration="13.806346967s" podCreationTimestamp="2025-11-24 08:33:02 +0000 UTC" firstStartedPulling="2025-11-24 08:33:04.84769767 +0000 UTC m=+1058.722842823" lastFinishedPulling="2025-11-24 08:33:13.660431644 +0000 UTC m=+1067.535576797" observedRunningTime="2025-11-24 08:33:15.804881946 +0000 UTC m=+1069.680027099" watchObservedRunningTime="2025-11-24 08:33:15.806346967 +0000 UTC m=+1069.681492120" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.848080 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-64c5f59669-7kdzn" podStartSLOduration=4.707711781 podStartE2EDuration="13.848061327s" podCreationTimestamp="2025-11-24 08:33:02 +0000 UTC" firstStartedPulling="2025-11-24 08:33:04.519137931 +0000 UTC m=+1058.394283084" lastFinishedPulling="2025-11-24 08:33:13.659487477 +0000 UTC m=+1067.534632630" observedRunningTime="2025-11-24 08:33:15.839493563 +0000 UTC m=+1069.714638736" watchObservedRunningTime="2025-11-24 08:33:15.848061327 +0000 UTC m=+1069.723206480" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.887033 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tqhj9" podStartSLOduration=4.645295001 podStartE2EDuration="48.887011918s" podCreationTimestamp="2025-11-24 08:32:27 +0000 UTC" firstStartedPulling="2025-11-24 08:32:29.428006912 +0000 UTC m=+1023.303152065" lastFinishedPulling="2025-11-24 08:33:13.669723829 +0000 UTC m=+1067.544868982" observedRunningTime="2025-11-24 08:33:15.877722163 +0000 UTC m=+1069.752867316" watchObservedRunningTime="2025-11-24 08:33:15.887011918 +0000 UTC m=+1069.762157071" Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.931720 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:33:15 crc kubenswrapper[4831]: I1124 08:33:15.942144 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-jstgx"] Nov 24 08:33:16 crc kubenswrapper[4831]: I1124 08:33:16.735804 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerID="373921e7da5f558ae1a83d765f0c0b5b4325185f57479d28b7a946dfe342a0f9" exitCode=2 Nov 24 08:33:16 crc kubenswrapper[4831]: I1124 08:33:16.735896 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerDied","Data":"373921e7da5f558ae1a83d765f0c0b5b4325185f57479d28b7a946dfe342a0f9"} Nov 24 08:33:16 crc kubenswrapper[4831]: I1124 08:33:16.968739 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59446fe2-a549-4797-8fbd-370904369dd4" path="/var/lib/kubelet/pods/59446fe2-a549-4797-8fbd-370904369dd4/volumes" Nov 24 08:33:17 crc kubenswrapper[4831]: I1124 08:33:17.620549 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:17 crc kubenswrapper[4831]: I1124 08:33:17.694550 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:18 crc kubenswrapper[4831]: I1124 08:33:18.454266 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:20 crc kubenswrapper[4831]: I1124 08:33:20.783049 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerID="62dda8189a1f996b74c360ba067e784d3798db44d290011dfd82101d743904d8" exitCode=0 Nov 24 08:33:20 crc kubenswrapper[4831]: I1124 08:33:20.783120 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerDied","Data":"62dda8189a1f996b74c360ba067e784d3798db44d290011dfd82101d743904d8"} Nov 24 08:33:20 crc kubenswrapper[4831]: I1124 08:33:20.895262 4831 scope.go:117] "RemoveContainer" containerID="a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae" Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.795167 4831 generic.go:334] "Generic (PLEG): container finished" podID="46a54561-e390-48c6-a27d-e516a013ebb3" containerID="2c1339dbdb125375189623fed665b0cf1f79e5068870958ef2c0c843ac1f510c" exitCode=0 Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.795327 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tqhj9" event={"ID":"46a54561-e390-48c6-a27d-e516a013ebb3","Type":"ContainerDied","Data":"2c1339dbdb125375189623fed665b0cf1f79e5068870958ef2c0c843ac1f510c"} Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.798091 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/2.log" Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.798989 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/1.log" Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.799892 4831 generic.go:334] "Generic (PLEG): container finished" podID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerID="e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da" exitCode=1 Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.799945 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerDied","Data":"e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da"} Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.800246 4831 scope.go:117] "RemoveContainer" containerID="a020b952b721dba6aaca6294c9250348d61480ae0ac84659ff5dca838ac420ae" Nov 24 08:33:21 crc kubenswrapper[4831]: I1124 08:33:21.801197 4831 scope.go:117] "RemoveContainer" containerID="e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da" Nov 24 08:33:21 crc kubenswrapper[4831]: E1124 08:33:21.801542 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-86987b4f86-lqghd_openstack(d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e)\"" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" Nov 24 08:33:22 crc kubenswrapper[4831]: I1124 08:33:22.814197 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/2.log" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.009061 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d4cd7dbc6-zgxq2" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.103505 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.103811 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f7b65778-ck4pk" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api-log" containerID="cri-o://5446ac08ad4e1d5be18e412e6b44963d742b99d385de1b4b1d2ab3138bcda82d" gracePeriod=30 Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.104533 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f7b65778-ck4pk" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api" containerID="cri-o://28a8488b69d05767971f9633c1d7b2c5493726941f0b04d6f975d59f2c8ce115" gracePeriod=30 Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.324960 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.482975 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483052 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483086 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483134 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcvtd\" (UniqueName: \"kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483165 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483193 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts\") pod \"46a54561-e390-48c6-a27d-e516a013ebb3\" (UID: \"46a54561-e390-48c6-a27d-e516a013ebb3\") " Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483218 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.483927 4831 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46a54561-e390-48c6-a27d-e516a013ebb3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.512241 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.514983 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts" (OuterVolumeSpecName: "scripts") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.516556 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd" (OuterVolumeSpecName: "kube-api-access-pcvtd") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "kube-api-access-pcvtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.518450 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.563461 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data" (OuterVolumeSpecName: "config-data") pod "46a54561-e390-48c6-a27d-e516a013ebb3" (UID: "46a54561-e390-48c6-a27d-e516a013ebb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.587393 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.587429 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.587441 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.587450 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcvtd\" (UniqueName: \"kubernetes.io/projected/46a54561-e390-48c6-a27d-e516a013ebb3-kube-api-access-pcvtd\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.587459 4831 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46a54561-e390-48c6-a27d-e516a013ebb3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.826364 4831 generic.go:334] "Generic (PLEG): container finished" podID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerID="5446ac08ad4e1d5be18e412e6b44963d742b99d385de1b4b1d2ab3138bcda82d" exitCode=143 Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.826375 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerDied","Data":"5446ac08ad4e1d5be18e412e6b44963d742b99d385de1b4b1d2ab3138bcda82d"} Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.828456 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tqhj9" event={"ID":"46a54561-e390-48c6-a27d-e516a013ebb3","Type":"ContainerDied","Data":"fb6a7d29b9ebfbef183ef07fb8c8b9816646431e7a3015127564cb3b19c3dc9f"} Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.828498 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb6a7d29b9ebfbef183ef07fb8c8b9816646431e7a3015127564cb3b19c3dc9f" Nov 24 08:33:23 crc kubenswrapper[4831]: I1124 08:33:23.828506 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tqhj9" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.166750 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:24 crc kubenswrapper[4831]: E1124 08:33:24.167516 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="dnsmasq-dns" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.167538 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="dnsmasq-dns" Nov 24 08:33:24 crc kubenswrapper[4831]: E1124 08:33:24.167569 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="init" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.167578 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="init" Nov 24 08:33:24 crc kubenswrapper[4831]: E1124 08:33:24.167606 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" containerName="cinder-db-sync" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.167615 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" containerName="cinder-db-sync" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.167803 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" containerName="cinder-db-sync" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.167836 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="59446fe2-a549-4797-8fbd-370904369dd4" containerName="dnsmasq-dns" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.173868 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.179474 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.179676 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.179834 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-b89k8" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.179980 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.184381 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.265381 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.271986 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.290608 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299075 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299133 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h7xr\" (UniqueName: \"kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299216 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299279 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299301 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.299346 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.320898 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.321742 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.323109 4831 scope.go:117] "RemoveContainer" containerID="e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da" Nov 24 08:33:24 crc kubenswrapper[4831]: E1124 08:33:24.323299 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-86987b4f86-lqghd_openstack(d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e)\"" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.343199 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-api" probeResult="failure" output="Get \"http://10.217.0.142:9696/\": dial tcp 10.217.0.142:9696: connect: connection refused" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400330 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h7xr\" (UniqueName: \"kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400388 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbsdt\" (UniqueName: \"kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400462 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400493 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400526 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400548 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400575 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400592 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400609 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400626 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400654 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.400900 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.405259 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.406837 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.406856 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.408770 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.433265 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h7xr\" (UniqueName: \"kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr\") pod \"cinder-scheduler-0\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.498145 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.501518 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.501673 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbsdt\" (UniqueName: \"kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.501811 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.501913 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.501987 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.503480 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.503519 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.504420 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.505140 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.506869 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.508349 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.511880 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.582329 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbsdt\" (UniqueName: \"kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt\") pod \"dnsmasq-dns-58db5546cc-k9rn6\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.603948 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.611524 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.611597 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.611649 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.612453 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.612515 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq4st\" (UniqueName: \"kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.612543 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.619337 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.645188 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.723856 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.723960 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.724023 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.724057 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.724088 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.724145 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.724178 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq4st\" (UniqueName: \"kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.726142 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.727758 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.741768 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.746410 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.747088 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.748493 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.752613 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq4st\" (UniqueName: \"kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st\") pod \"cinder-api-0\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " pod="openstack/cinder-api-0" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.842512 4831 scope.go:117] "RemoveContainer" containerID="e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da" Nov 24 08:33:24 crc kubenswrapper[4831]: E1124 08:33:24.842706 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"neutron-httpd\" with CrashLoopBackOff: \"back-off 20s restarting failed container=neutron-httpd pod=neutron-86987b4f86-lqghd_openstack(d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e)\"" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" Nov 24 08:33:24 crc kubenswrapper[4831]: I1124 08:33:24.931205 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.163267 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.284312 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:33:25 crc kubenswrapper[4831]: W1124 08:33:25.287386 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfe5b851_71ed_43f4_b92a_2a3a913686b8.slice/crio-887875c472deb79b95b02e36292f298a2a1bde9f8b5e865f35e48898580df060 WatchSource:0}: Error finding container 887875c472deb79b95b02e36292f298a2a1bde9f8b5e865f35e48898580df060: Status 404 returned error can't find the container with id 887875c472deb79b95b02e36292f298a2a1bde9f8b5e865f35e48898580df060 Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.496264 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.860507 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerStarted","Data":"7cdbb1ef23f67c6f0aea6b2dfc0acfc637ccb1fc61d6156f1fa9dc872a9c43a9"} Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.862551 4831 generic.go:334] "Generic (PLEG): container finished" podID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerID="d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855" exitCode=0 Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.863601 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" event={"ID":"cfe5b851-71ed-43f4-b92a-2a3a913686b8","Type":"ContainerDied","Data":"d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855"} Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.863638 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" event={"ID":"cfe5b851-71ed-43f4-b92a-2a3a913686b8","Type":"ContainerStarted","Data":"887875c472deb79b95b02e36292f298a2a1bde9f8b5e865f35e48898580df060"} Nov 24 08:33:25 crc kubenswrapper[4831]: I1124 08:33:25.866548 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerStarted","Data":"a26b7ddc5fe9c50d6ac7acf8ef4890dac88fb2bce3f487bdf92898e11f68ca9f"} Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.326217 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7b65778-ck4pk" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:35266->10.217.0.149:9311: read: connection reset by peer" Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.327879 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5f7b65778-ck4pk" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.149:9311/healthcheck\": read tcp 10.217.0.2:35274->10.217.0.149:9311: read: connection reset by peer" Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.682310 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.919991 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerStarted","Data":"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d"} Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.931126 4831 generic.go:334] "Generic (PLEG): container finished" podID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerID="28a8488b69d05767971f9633c1d7b2c5493726941f0b04d6f975d59f2c8ce115" exitCode=0 Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.931302 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerDied","Data":"28a8488b69d05767971f9633c1d7b2c5493726941f0b04d6f975d59f2c8ce115"} Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.936576 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" event={"ID":"cfe5b851-71ed-43f4-b92a-2a3a913686b8","Type":"ContainerStarted","Data":"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0"} Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.937653 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.953751 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerStarted","Data":"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd"} Nov 24 08:33:26 crc kubenswrapper[4831]: I1124 08:33:26.958988 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.011410 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data\") pod \"cbc69059-ddcf-4352-803c-372e76f8f50d\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.011696 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom\") pod \"cbc69059-ddcf-4352-803c-372e76f8f50d\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.011732 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs\") pod \"cbc69059-ddcf-4352-803c-372e76f8f50d\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.011750 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle\") pod \"cbc69059-ddcf-4352-803c-372e76f8f50d\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.011944 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2rrp\" (UniqueName: \"kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp\") pod \"cbc69059-ddcf-4352-803c-372e76f8f50d\" (UID: \"cbc69059-ddcf-4352-803c-372e76f8f50d\") " Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.025688 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs" (OuterVolumeSpecName: "logs") pod "cbc69059-ddcf-4352-803c-372e76f8f50d" (UID: "cbc69059-ddcf-4352-803c-372e76f8f50d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.031119 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cbc69059-ddcf-4352-803c-372e76f8f50d" (UID: "cbc69059-ddcf-4352-803c-372e76f8f50d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.052219 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp" (OuterVolumeSpecName: "kube-api-access-v2rrp") pod "cbc69059-ddcf-4352-803c-372e76f8f50d" (UID: "cbc69059-ddcf-4352-803c-372e76f8f50d"). InnerVolumeSpecName "kube-api-access-v2rrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.140039 4831 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.140069 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbc69059-ddcf-4352-803c-372e76f8f50d-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.140079 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2rrp\" (UniqueName: \"kubernetes.io/projected/cbc69059-ddcf-4352-803c-372e76f8f50d-kube-api-access-v2rrp\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.142650 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data" (OuterVolumeSpecName: "config-data") pod "cbc69059-ddcf-4352-803c-372e76f8f50d" (UID: "cbc69059-ddcf-4352-803c-372e76f8f50d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.147330 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbc69059-ddcf-4352-803c-372e76f8f50d" (UID: "cbc69059-ddcf-4352-803c-372e76f8f50d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.163065 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" podStartSLOduration=3.163042436 podStartE2EDuration="3.163042436s" podCreationTimestamp="2025-11-24 08:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:27.042074346 +0000 UTC m=+1080.917219519" watchObservedRunningTime="2025-11-24 08:33:27.163042436 +0000 UTC m=+1081.038187589" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.241455 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.241500 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbc69059-ddcf-4352-803c-372e76f8f50d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.964054 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerStarted","Data":"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b"} Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.964201 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api-log" containerID="cri-o://8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" gracePeriod=30 Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.964312 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api" containerID="cri-o://53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" gracePeriod=30 Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.964543 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.967095 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f7b65778-ck4pk" event={"ID":"cbc69059-ddcf-4352-803c-372e76f8f50d","Type":"ContainerDied","Data":"f369f6553f77d6afb94446412b62134da4985825bc18323168dec2f5fc230f7a"} Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.967128 4831 scope.go:117] "RemoveContainer" containerID="28a8488b69d05767971f9633c1d7b2c5493726941f0b04d6f975d59f2c8ce115" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.967248 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f7b65778-ck4pk" Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.977936 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerStarted","Data":"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e"} Nov 24 08:33:27 crc kubenswrapper[4831]: I1124 08:33:27.993839 4831 scope.go:117] "RemoveContainer" containerID="5446ac08ad4e1d5be18e412e6b44963d742b99d385de1b4b1d2ab3138bcda82d" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.015435 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.015411721 podStartE2EDuration="4.015411721s" podCreationTimestamp="2025-11-24 08:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:27.992453196 +0000 UTC m=+1081.867598359" watchObservedRunningTime="2025-11-24 08:33:28.015411721 +0000 UTC m=+1081.890556874" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.017971 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.030507 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f7b65778-ck4pk"] Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.039328 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.232320831 podStartE2EDuration="4.039290022s" podCreationTimestamp="2025-11-24 08:33:24 +0000 UTC" firstStartedPulling="2025-11-24 08:33:25.17484915 +0000 UTC m=+1079.049994303" lastFinishedPulling="2025-11-24 08:33:25.981818341 +0000 UTC m=+1079.856963494" observedRunningTime="2025-11-24 08:33:28.037876051 +0000 UTC m=+1081.913021214" watchObservedRunningTime="2025-11-24 08:33:28.039290022 +0000 UTC m=+1081.914435175" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.280408 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79dd578589-cbqvb" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.318541 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.368849 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.369068 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-86987b4f86-lqghd" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-api" containerID="cri-o://d529b16349b5d6a3383d438f6bfebb9abb75bbe26ecf56ffeff69709f5810db5" gracePeriod=30 Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.404845 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.405025 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.603594 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670272 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670421 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670497 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq4st\" (UniqueName: \"kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670501 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670529 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670576 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670654 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670688 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data\") pod \"6e1e0e25-7249-4c75-8204-43e2842ed81a\" (UID: \"6e1e0e25-7249-4c75-8204-43e2842ed81a\") " Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.670974 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs" (OuterVolumeSpecName: "logs") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.671398 4831 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e1e0e25-7249-4c75-8204-43e2842ed81a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.671418 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e1e0e25-7249-4c75-8204-43e2842ed81a-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.711141 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st" (OuterVolumeSpecName: "kube-api-access-cq4st") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "kube-api-access-cq4st". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.711448 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts" (OuterVolumeSpecName: "scripts") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.715989 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.727618 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.774681 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq4st\" (UniqueName: \"kubernetes.io/projected/6e1e0e25-7249-4c75-8204-43e2842ed81a-kube-api-access-cq4st\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.774867 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.774922 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.774976 4831 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.777449 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data" (OuterVolumeSpecName: "config-data") pod "6e1e0e25-7249-4c75-8204-43e2842ed81a" (UID: "6e1e0e25-7249-4c75-8204-43e2842ed81a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.876411 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e1e0e25-7249-4c75-8204-43e2842ed81a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.918956 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" path="/var/lib/kubelet/pods/cbc69059-ddcf-4352-803c-372e76f8f50d/volumes" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.989400 4831 generic.go:334] "Generic (PLEG): container finished" podID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerID="53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" exitCode=0 Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.989825 4831 generic.go:334] "Generic (PLEG): container finished" podID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerID="8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" exitCode=143 Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.989463 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerDied","Data":"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b"} Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.990458 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerDied","Data":"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d"} Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.989446 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.990476 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6e1e0e25-7249-4c75-8204-43e2842ed81a","Type":"ContainerDied","Data":"7cdbb1ef23f67c6f0aea6b2dfc0acfc637ccb1fc61d6156f1fa9dc872a9c43a9"} Nov 24 08:33:28 crc kubenswrapper[4831]: I1124 08:33:28.990496 4831 scope.go:117] "RemoveContainer" containerID="53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.012477 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.013890 4831 scope.go:117] "RemoveContainer" containerID="8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.022508 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.033620 4831 scope.go:117] "RemoveContainer" containerID="53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.035487 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b\": container with ID starting with 53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b not found: ID does not exist" containerID="53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.035551 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b"} err="failed to get container status \"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b\": rpc error: code = NotFound desc = could not find container \"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b\": container with ID starting with 53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b not found: ID does not exist" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.035584 4831 scope.go:117] "RemoveContainer" containerID="8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.038841 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d\": container with ID starting with 8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d not found: ID does not exist" containerID="8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.038891 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d"} err="failed to get container status \"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d\": rpc error: code = NotFound desc = could not find container \"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d\": container with ID starting with 8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d not found: ID does not exist" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.038946 4831 scope.go:117] "RemoveContainer" containerID="53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.039258 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b"} err="failed to get container status \"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b\": rpc error: code = NotFound desc = could not find container \"53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b\": container with ID starting with 53212fe47841eb020dc46ef084a79d62c84c8dee25e781bf4bca0af51490221b not found: ID does not exist" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.039303 4831 scope.go:117] "RemoveContainer" containerID="8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.039543 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d"} err="failed to get container status \"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d\": rpc error: code = NotFound desc = could not find container \"8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d\": container with ID starting with 8be15d76f3eed972abc1e543249cdfc1c00113fbf95910502d3210029077fe9d not found: ID does not exist" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.073979 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.074419 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074444 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api" Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.074464 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074472 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api" Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.074493 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074501 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: E1124 08:33:29.074517 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074525 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074702 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074721 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbc69059-ddcf-4352-803c-372e76f8f50d" containerName="barbican-api" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074733 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api-log" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.074744 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" containerName="cinder-api" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.075677 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.080995 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.081146 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.081260 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.095739 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185436 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-scripts\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185486 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7795091a-65df-4971-b520-e9cdaa870328-logs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185508 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185642 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185677 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7795091a-65df-4971-b520-e9cdaa870328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185764 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185781 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data-custom\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185798 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhz26\" (UniqueName: \"kubernetes.io/projected/7795091a-65df-4971-b520-e9cdaa870328-kube-api-access-dhz26\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.185878 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287290 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7795091a-65df-4971-b520-e9cdaa870328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287389 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7795091a-65df-4971-b520-e9cdaa870328-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287417 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287443 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data-custom\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287464 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhz26\" (UniqueName: \"kubernetes.io/projected/7795091a-65df-4971-b520-e9cdaa870328-kube-api-access-dhz26\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287526 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287573 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-scripts\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.287591 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7795091a-65df-4971-b520-e9cdaa870328-logs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.288118 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7795091a-65df-4971-b520-e9cdaa870328-logs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.288180 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.288241 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.291692 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.291801 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.294009 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.294954 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-public-tls-certs\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.295133 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-config-data-custom\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.295824 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7795091a-65df-4971-b520-e9cdaa870328-scripts\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.307978 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhz26\" (UniqueName: \"kubernetes.io/projected/7795091a-65df-4971-b520-e9cdaa870328-kube-api-access-dhz26\") pod \"cinder-api-0\" (UID: \"7795091a-65df-4971-b520-e9cdaa870328\") " pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.394063 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.499437 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 08:33:29 crc kubenswrapper[4831]: I1124 08:33:29.910903 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 24 08:33:30 crc kubenswrapper[4831]: I1124 08:33:30.008838 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7795091a-65df-4971-b520-e9cdaa870328","Type":"ContainerStarted","Data":"01cf06c581d4d815b472c1cfe6a6e78ba73279c1a0cdb06f80b3f7a1ff53146b"} Nov 24 08:33:30 crc kubenswrapper[4831]: I1124 08:33:30.904851 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e1e0e25-7249-4c75-8204-43e2842ed81a" path="/var/lib/kubelet/pods/6e1e0e25-7249-4c75-8204-43e2842ed81a/volumes" Nov 24 08:33:31 crc kubenswrapper[4831]: I1124 08:33:31.022975 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7795091a-65df-4971-b520-e9cdaa870328","Type":"ContainerStarted","Data":"2a88d38f06cafe6c3f361711b7ba18615121dee07246f2d18b9594e098285c5d"} Nov 24 08:33:32 crc kubenswrapper[4831]: I1124 08:33:32.037715 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7795091a-65df-4971-b520-e9cdaa870328","Type":"ContainerStarted","Data":"8484522bd23e2d34cfacdf646c3ca2aaef45b138e43b96f532c12868e3789beb"} Nov 24 08:33:32 crc kubenswrapper[4831]: I1124 08:33:32.037900 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 24 08:33:32 crc kubenswrapper[4831]: I1124 08:33:32.067848 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.06782671 podStartE2EDuration="3.06782671s" podCreationTimestamp="2025-11-24 08:33:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:32.064486735 +0000 UTC m=+1085.939631908" watchObservedRunningTime="2025-11-24 08:33:32.06782671 +0000 UTC m=+1085.942971863" Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.422555 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-77b5948bf-z797k" Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.609601 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.772770 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.772987 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="dnsmasq-dns" containerID="cri-o://02892049a4e10c5a353df4ae463fac74d7a38b64ab923c6b87dbce7be0772d85" gracePeriod=10 Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.908648 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 08:33:34 crc kubenswrapper[4831]: I1124 08:33:34.977256 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.092599 4831 generic.go:334] "Generic (PLEG): container finished" podID="821e46e4-876a-4b97-b4be-70baecc6b883" containerID="02892049a4e10c5a353df4ae463fac74d7a38b64ab923c6b87dbce7be0772d85" exitCode=0 Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.092678 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" event={"ID":"821e46e4-876a-4b97-b4be-70baecc6b883","Type":"ContainerDied","Data":"02892049a4e10c5a353df4ae463fac74d7a38b64ab923c6b87dbce7be0772d85"} Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.093173 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="cinder-scheduler" containerID="cri-o://63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd" gracePeriod=30 Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.093425 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="probe" containerID="cri-o://7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e" gracePeriod=30 Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.389520 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.549631 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc\") pod \"821e46e4-876a-4b97-b4be-70baecc6b883\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.549812 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb\") pod \"821e46e4-876a-4b97-b4be-70baecc6b883\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.549837 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb\") pod \"821e46e4-876a-4b97-b4be-70baecc6b883\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.549882 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9swmv\" (UniqueName: \"kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv\") pod \"821e46e4-876a-4b97-b4be-70baecc6b883\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.549902 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config\") pod \"821e46e4-876a-4b97-b4be-70baecc6b883\" (UID: \"821e46e4-876a-4b97-b4be-70baecc6b883\") " Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.570089 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv" (OuterVolumeSpecName: "kube-api-access-9swmv") pod "821e46e4-876a-4b97-b4be-70baecc6b883" (UID: "821e46e4-876a-4b97-b4be-70baecc6b883"). InnerVolumeSpecName "kube-api-access-9swmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.605128 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.654455 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9swmv\" (UniqueName: \"kubernetes.io/projected/821e46e4-876a-4b97-b4be-70baecc6b883-kube-api-access-9swmv\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.664403 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "821e46e4-876a-4b97-b4be-70baecc6b883" (UID: "821e46e4-876a-4b97-b4be-70baecc6b883"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.675190 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config" (OuterVolumeSpecName: "config") pod "821e46e4-876a-4b97-b4be-70baecc6b883" (UID: "821e46e4-876a-4b97-b4be-70baecc6b883"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.676289 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "821e46e4-876a-4b97-b4be-70baecc6b883" (UID: "821e46e4-876a-4b97-b4be-70baecc6b883"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.685634 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "821e46e4-876a-4b97-b4be-70baecc6b883" (UID: "821e46e4-876a-4b97-b4be-70baecc6b883"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.686820 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6f67674df6-sk4ws" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.758755 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.758788 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.758799 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:35 crc kubenswrapper[4831]: I1124 08:33:35.758809 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/821e46e4-876a-4b97-b4be-70baecc6b883-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.107471 4831 generic.go:334] "Generic (PLEG): container finished" podID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerID="7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e" exitCode=0 Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.107561 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerDied","Data":"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e"} Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.110277 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" event={"ID":"821e46e4-876a-4b97-b4be-70baecc6b883","Type":"ContainerDied","Data":"4ecf8a4f87efb05d402a4d09b2ca17cdbc44c62ff79fbf6e76aa98302efe3b35"} Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.110345 4831 scope.go:117] "RemoveContainer" containerID="02892049a4e10c5a353df4ae463fac74d7a38b64ab923c6b87dbce7be0772d85" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.110287 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-cmjxr" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.138541 4831 scope.go:117] "RemoveContainer" containerID="6d7e84c26d3fcf180c0c88849374e1fd9aad2d2a6a76c0aa84cedbd744e10322" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.145738 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.152686 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-cmjxr"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.421585 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: E1124 08:33:36.421932 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="dnsmasq-dns" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.421948 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="dnsmasq-dns" Nov 24 08:33:36 crc kubenswrapper[4831]: E1124 08:33:36.421981 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="init" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.421988 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="init" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.422166 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" containerName="dnsmasq-dns" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.422749 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.425426 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.426109 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-f4wl6" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.426265 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.441461 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.571746 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.571831 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.571868 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtdcf\" (UniqueName: \"kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.571936 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.673698 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.673811 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.673860 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.673885 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtdcf\" (UniqueName: \"kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.674873 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.687125 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.688688 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.705234 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtdcf\" (UniqueName: \"kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf\") pod \"openstackclient\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.709908 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.710586 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.735138 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.824168 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.829162 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.853908 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.882062 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config-secret\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.882287 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.882429 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.882516 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56fbm\" (UniqueName: \"kubernetes.io/projected/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-kube-api-access-56fbm\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.908588 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="821e46e4-876a-4b97-b4be-70baecc6b883" path="/var/lib/kubelet/pods/821e46e4-876a-4b97-b4be-70baecc6b883/volumes" Nov 24 08:33:36 crc kubenswrapper[4831]: E1124 08:33:36.920774 4831 log.go:32] "RunPodSandbox from runtime service failed" err=< Nov 24 08:33:36 crc kubenswrapper[4831]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8667bb63-2e26-4798-86aa-715a527bf256_0(bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd" Netns:"/var/run/netns/ad6b084c-c2f6-439a-97c3-1ff1773ccccf" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd;K8S_POD_UID=8667bb63-2e26-4798-86aa-715a527bf256" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8667bb63-2e26-4798-86aa-715a527bf256]: expected pod UID "8667bb63-2e26-4798-86aa-715a527bf256" but got "8db8fec2-ed06-4ad5-8fcf-a491ff18abd1" from Kube API Nov 24 08:33:36 crc kubenswrapper[4831]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 08:33:36 crc kubenswrapper[4831]: > Nov 24 08:33:36 crc kubenswrapper[4831]: E1124 08:33:36.920884 4831 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Nov 24 08:33:36 crc kubenswrapper[4831]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8667bb63-2e26-4798-86aa-715a527bf256_0(bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd" Netns:"/var/run/netns/ad6b084c-c2f6-439a-97c3-1ff1773ccccf" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=bca5b220421bfcd02aa107d7be8930892bf8e135fd6dffc6e84725922cd0a7fd;K8S_POD_UID=8667bb63-2e26-4798-86aa-715a527bf256" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8667bb63-2e26-4798-86aa-715a527bf256]: expected pod UID "8667bb63-2e26-4798-86aa-715a527bf256" but got "8db8fec2-ed06-4ad5-8fcf-a491ff18abd1" from Kube API Nov 24 08:33:36 crc kubenswrapper[4831]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Nov 24 08:33:36 crc kubenswrapper[4831]: > pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.983737 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config-secret\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.983811 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.983860 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.983907 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56fbm\" (UniqueName: \"kubernetes.io/projected/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-kube-api-access-56fbm\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:36 crc kubenswrapper[4831]: I1124 08:33:36.985856 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.001971 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-openstack-config-secret\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.008215 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.008816 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56fbm\" (UniqueName: \"kubernetes.io/projected/8db8fec2-ed06-4ad5-8fcf-a491ff18abd1-kube-api-access-56fbm\") pod \"openstackclient\" (UID: \"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1\") " pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.120432 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.133097 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.135222 4831 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8667bb63-2e26-4798-86aa-715a527bf256" podUID="8db8fec2-ed06-4ad5-8fcf-a491ff18abd1" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.187338 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret\") pod \"8667bb63-2e26-4798-86aa-715a527bf256\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.187391 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle\") pod \"8667bb63-2e26-4798-86aa-715a527bf256\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.187477 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtdcf\" (UniqueName: \"kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf\") pod \"8667bb63-2e26-4798-86aa-715a527bf256\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.187631 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config\") pod \"8667bb63-2e26-4798-86aa-715a527bf256\" (UID: \"8667bb63-2e26-4798-86aa-715a527bf256\") " Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.188275 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8667bb63-2e26-4798-86aa-715a527bf256" (UID: "8667bb63-2e26-4798-86aa-715a527bf256"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.191200 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8667bb63-2e26-4798-86aa-715a527bf256" (UID: "8667bb63-2e26-4798-86aa-715a527bf256"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.192894 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8667bb63-2e26-4798-86aa-715a527bf256" (UID: "8667bb63-2e26-4798-86aa-715a527bf256"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.215601 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf" (OuterVolumeSpecName: "kube-api-access-mtdcf") pod "8667bb63-2e26-4798-86aa-715a527bf256" (UID: "8667bb63-2e26-4798-86aa-715a527bf256"). InnerVolumeSpecName "kube-api-access-mtdcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.289660 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtdcf\" (UniqueName: \"kubernetes.io/projected/8667bb63-2e26-4798-86aa-715a527bf256-kube-api-access-mtdcf\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.290027 4831 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.290040 4831 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.290051 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8667bb63-2e26-4798-86aa-715a527bf256-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.291059 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:37 crc kubenswrapper[4831]: I1124 08:33:37.818259 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 24 08:33:38 crc kubenswrapper[4831]: I1124 08:33:38.127911 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 24 08:33:38 crc kubenswrapper[4831]: I1124 08:33:38.132436 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1","Type":"ContainerStarted","Data":"8ed7b220077c276979b46463f8a6394fdf5639b38a349982e68867f85c1f2474"} Nov 24 08:33:38 crc kubenswrapper[4831]: I1124 08:33:38.145363 4831 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8667bb63-2e26-4798-86aa-715a527bf256" podUID="8db8fec2-ed06-4ad5-8fcf-a491ff18abd1" Nov 24 08:33:38 crc kubenswrapper[4831]: I1124 08:33:38.919417 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8667bb63-2e26-4798-86aa-715a527bf256" path="/var/lib/kubelet/pods/8667bb63-2e26-4798-86aa-715a527bf256/volumes" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.036013 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.138845 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.138952 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.139008 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.139058 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.139104 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h7xr\" (UniqueName: \"kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.139132 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data\") pod \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\" (UID: \"5447c7e4-c964-4b43-a3f9-d33345f4ae15\") " Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.140709 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.150471 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.161687 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts" (OuterVolumeSpecName: "scripts") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.161855 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr" (OuterVolumeSpecName: "kube-api-access-7h7xr") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "kube-api-access-7h7xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.184789 4831 generic.go:334] "Generic (PLEG): container finished" podID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerID="63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd" exitCode=0 Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.184836 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerDied","Data":"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd"} Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.184864 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5447c7e4-c964-4b43-a3f9-d33345f4ae15","Type":"ContainerDied","Data":"a26b7ddc5fe9c50d6ac7acf8ef4890dac88fb2bce3f487bdf92898e11f68ca9f"} Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.184882 4831 scope.go:117] "RemoveContainer" containerID="7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.185055 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.243155 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.243360 4831 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5447c7e4-c964-4b43-a3f9-d33345f4ae15-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.243441 4831 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.243571 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h7xr\" (UniqueName: \"kubernetes.io/projected/5447c7e4-c964-4b43-a3f9-d33345f4ae15-kube-api-access-7h7xr\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.273550 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.274561 4831 scope.go:117] "RemoveContainer" containerID="63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.298874 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data" (OuterVolumeSpecName: "config-data") pod "5447c7e4-c964-4b43-a3f9-d33345f4ae15" (UID: "5447c7e4-c964-4b43-a3f9-d33345f4ae15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.318738 4831 scope.go:117] "RemoveContainer" containerID="7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e" Nov 24 08:33:40 crc kubenswrapper[4831]: E1124 08:33:40.320443 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e\": container with ID starting with 7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e not found: ID does not exist" containerID="7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.320486 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e"} err="failed to get container status \"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e\": rpc error: code = NotFound desc = could not find container \"7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e\": container with ID starting with 7a47804bccc7690f16abcdc773c1251576dfbe962e6967ee2aff743eea07dd2e not found: ID does not exist" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.320514 4831 scope.go:117] "RemoveContainer" containerID="63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd" Nov 24 08:33:40 crc kubenswrapper[4831]: E1124 08:33:40.320877 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd\": container with ID starting with 63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd not found: ID does not exist" containerID="63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.320904 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd"} err="failed to get container status \"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd\": rpc error: code = NotFound desc = could not find container \"63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd\": container with ID starting with 63b7c5a97a7c5e1b4561fe008e63e15ab5fec456e0964e0861a4504b3fee2bbd not found: ID does not exist" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.345038 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.345093 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5447c7e4-c964-4b43-a3f9-d33345f4ae15-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.515268 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.525627 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.547776 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:40 crc kubenswrapper[4831]: E1124 08:33:40.552051 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="cinder-scheduler" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.552079 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="cinder-scheduler" Nov 24 08:33:40 crc kubenswrapper[4831]: E1124 08:33:40.552104 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="probe" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.552111 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="probe" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.552386 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="cinder-scheduler" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.552409 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" containerName="probe" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.553800 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.557423 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.619980 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656017 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656106 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656168 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-scripts\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656230 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656271 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.656468 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-844gv\" (UniqueName: \"kubernetes.io/projected/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-kube-api-access-844gv\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.759059 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.759172 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-scripts\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.760070 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.760149 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.760191 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.760547 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-844gv\" (UniqueName: \"kubernetes.io/projected/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-kube-api-access-844gv\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.760616 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.767526 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.768465 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-scripts\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.776042 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-config-data\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.785158 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.799487 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-844gv\" (UniqueName: \"kubernetes.io/projected/c13cfbd9-7d6f-4c2d-96aa-5de94bda5352-kube-api-access-844gv\") pod \"cinder-scheduler-0\" (UID: \"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352\") " pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.874786 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 24 08:33:40 crc kubenswrapper[4831]: I1124 08:33:40.909884 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5447c7e4-c964-4b43-a3f9-d33345f4ae15" path="/var/lib/kubelet/pods/5447c7e4-c964-4b43-a3f9-d33345f4ae15/volumes" Nov 24 08:33:41 crc kubenswrapper[4831]: I1124 08:33:41.358419 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 24 08:33:41 crc kubenswrapper[4831]: W1124 08:33:41.364865 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc13cfbd9_7d6f_4c2d_96aa_5de94bda5352.slice/crio-4752be4a3f407e59512e808b6c5692602cefa578ba165c6f722a1b8487de9a13 WatchSource:0}: Error finding container 4752be4a3f407e59512e808b6c5692602cefa578ba165c6f722a1b8487de9a13: Status 404 returned error can't find the container with id 4752be4a3f407e59512e808b6c5692602cefa578ba165c6f722a1b8487de9a13 Nov 24 08:33:42 crc kubenswrapper[4831]: I1124 08:33:42.116509 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 24 08:33:42 crc kubenswrapper[4831]: I1124 08:33:42.256232 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352","Type":"ContainerStarted","Data":"d42eb4a0e46d7effaa3c54ae9a5c72093484106e7392ddde5676ae3ed15c32d2"} Nov 24 08:33:42 crc kubenswrapper[4831]: I1124 08:33:42.256378 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352","Type":"ContainerStarted","Data":"4752be4a3f407e59512e808b6c5692602cefa578ba165c6f722a1b8487de9a13"} Nov 24 08:33:43 crc kubenswrapper[4831]: I1124 08:33:43.268947 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c13cfbd9-7d6f-4c2d-96aa-5de94bda5352","Type":"ContainerStarted","Data":"a09c52d41fb7152dcae2af1510854931fdb39aeb964a14349215fd055aa91710"} Nov 24 08:33:43 crc kubenswrapper[4831]: I1124 08:33:43.286079 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.28605838 podStartE2EDuration="3.28605838s" podCreationTimestamp="2025-11-24 08:33:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:43.28396229 +0000 UTC m=+1097.159107453" watchObservedRunningTime="2025-11-24 08:33:43.28605838 +0000 UTC m=+1097.161203523" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.281309 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/2.log" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.282497 4831 generic.go:334] "Generic (PLEG): container finished" podID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerID="d529b16349b5d6a3383d438f6bfebb9abb75bbe26ecf56ffeff69709f5810db5" exitCode=0 Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.283301 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerDied","Data":"d529b16349b5d6a3383d438f6bfebb9abb75bbe26ecf56ffeff69709f5810db5"} Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.477279 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/2.log" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.479275 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.550179 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle\") pod \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.550291 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fds5k\" (UniqueName: \"kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k\") pod \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.550451 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config\") pod \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.550487 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs\") pod \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.550584 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config\") pod \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\" (UID: \"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e\") " Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.558952 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" (UID: "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.560151 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k" (OuterVolumeSpecName: "kube-api-access-fds5k") pod "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" (UID: "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e"). InnerVolumeSpecName "kube-api-access-fds5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.652572 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fds5k\" (UniqueName: \"kubernetes.io/projected/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-kube-api-access-fds5k\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.652603 4831 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.662460 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config" (OuterVolumeSpecName: "config") pod "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" (UID: "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.682569 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" (UID: "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.709420 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" (UID: "d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.754041 4831 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.754072 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:44 crc kubenswrapper[4831]: I1124 08:33:44.754081 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.293433 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-86987b4f86-lqghd_d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/neutron-httpd/2.log" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.294137 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-86987b4f86-lqghd" event={"ID":"d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e","Type":"ContainerDied","Data":"61ed820e2070f79c2ce479c4bcea659a7152f8475c86da3df899e6c05865a7de"} Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.294184 4831 scope.go:117] "RemoveContainer" containerID="e532a382586aeab0ef379124d0202c9eef4248a2561351dc6dd5e591644818da" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.294301 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-86987b4f86-lqghd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.313204 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.323964 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-86987b4f86-lqghd"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.335599 4831 scope.go:117] "RemoveContainer" containerID="d529b16349b5d6a3383d438f6bfebb9abb75bbe26ecf56ffeff69709f5810db5" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.688688 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5z8s4"] Nov 24 08:33:45 crc kubenswrapper[4831]: E1124 08:33:45.689166 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689192 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: E1124 08:33:45.689213 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689221 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: E1124 08:33:45.689239 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-api" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689247 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-api" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689475 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689487 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.689505 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-api" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.690211 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.711930 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5z8s4"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.769357 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc74r\" (UniqueName: \"kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.770071 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.789687 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-zdfb7"] Nov 24 08:33:45 crc kubenswrapper[4831]: E1124 08:33:45.790034 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.790049 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.790232 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" containerName="neutron-httpd" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.790789 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.800940 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zdfb7"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.872342 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.872437 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.872483 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc74r\" (UniqueName: \"kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.872516 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcxfs\" (UniqueName: \"kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.873572 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.875762 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.911272 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-8vhg7"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.915449 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc74r\" (UniqueName: \"kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r\") pod \"nova-api-db-create-5z8s4\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.915717 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.962392 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1900-account-create-z9bh8"] Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.963658 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.975672 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcxfs\" (UniqueName: \"kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.975803 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.976570 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.977235 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 24 08:33:45 crc kubenswrapper[4831]: I1124 08:33:45.992584 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8vhg7"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.010168 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1900-account-create-z9bh8"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.011255 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.013790 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcxfs\" (UniqueName: \"kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs\") pod \"nova-cell0-db-create-zdfb7\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.078903 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.078967 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt9ln\" (UniqueName: \"kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.079030 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.079195 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqfls\" (UniqueName: \"kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.131158 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-8f65-account-create-96zkh"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.132423 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.137673 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.138978 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.166308 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8f65-account-create-96zkh"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.180280 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.180534 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt9ln\" (UniqueName: \"kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.180632 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.180752 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqfls\" (UniqueName: \"kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.181751 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.182414 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.204834 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt9ln\" (UniqueName: \"kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln\") pod \"nova-cell1-db-create-8vhg7\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.214336 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqfls\" (UniqueName: \"kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls\") pod \"nova-api-1900-account-create-z9bh8\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.282364 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.282701 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8lxf\" (UniqueName: \"kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.298552 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ed31-account-create-fw7hs"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.299987 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.302696 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.304752 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.309604 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ed31-account-create-fw7hs"] Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.320118 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerID="d71f8a55572711739f001d9af6310dd8a6baeb6d675851090dca851564493f9e" exitCode=137 Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.320168 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerDied","Data":"d71f8a55572711739f001d9af6310dd8a6baeb6d675851090dca851564493f9e"} Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.358693 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.387546 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8lxf\" (UniqueName: \"kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.387799 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.389509 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.407510 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8lxf\" (UniqueName: \"kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf\") pod \"nova-cell0-8f65-account-create-96zkh\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.475257 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.491979 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.492283 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp496\" (UniqueName: \"kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.594677 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp496\" (UniqueName: \"kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.594805 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.595539 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.612368 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp496\" (UniqueName: \"kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496\") pod \"nova-cell1-ed31-account-create-fw7hs\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.619475 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:46 crc kubenswrapper[4831]: I1124 08:33:46.908987 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e" path="/var/lib/kubelet/pods/d9cde5cb-8b3b-43f6-a7f1-49ee4f25865e/volumes" Nov 24 08:33:51 crc kubenswrapper[4831]: I1124 08:33:51.135785 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.400520 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.426215 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b","Type":"ContainerDied","Data":"78959f21897d720ef11a9f0a8e0739ab784a02f478fc45a2eed9a0c7d60509ee"} Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.427187 4831 scope.go:117] "RemoveContainer" containerID="d71f8a55572711739f001d9af6310dd8a6baeb6d675851090dca851564493f9e" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.478416 4831 scope.go:117] "RemoveContainer" containerID="373921e7da5f558ae1a83d765f0c0b5b4325185f57479d28b7a946dfe342a0f9" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.515130 4831 scope.go:117] "RemoveContainer" containerID="62dda8189a1f996b74c360ba067e784d3798db44d290011dfd82101d743904d8" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.532833 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq2d9\" (UniqueName: \"kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533062 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533177 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533206 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533236 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533267 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.533300 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd\") pod \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\" (UID: \"5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b\") " Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.534209 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.535349 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.548831 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9" (OuterVolumeSpecName: "kube-api-access-gq2d9") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "kube-api-access-gq2d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.548999 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts" (OuterVolumeSpecName: "scripts") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: W1124 08:33:53.561713 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb639f94c_6740_497d_b53e_4c297f65d750.slice/crio-11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8 WatchSource:0}: Error finding container 11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8: Status 404 returned error can't find the container with id 11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8 Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.584339 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.593297 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8vhg7"] Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.599807 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636047 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636577 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636592 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636604 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636617 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.636627 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq2d9\" (UniqueName: \"kubernetes.io/projected/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-kube-api-access-gq2d9\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.658661 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data" (OuterVolumeSpecName: "config-data") pod "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" (UID: "5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.739732 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.796529 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-8f65-account-create-96zkh"] Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.805359 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ed31-account-create-fw7hs"] Nov 24 08:33:53 crc kubenswrapper[4831]: W1124 08:33:53.809807 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fb1dd7b_8f0b_4d59_82a8_58e5551b9cfe.slice/crio-fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea WatchSource:0}: Error finding container fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea: Status 404 returned error can't find the container with id fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.820810 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zdfb7"] Nov 24 08:33:53 crc kubenswrapper[4831]: I1124 08:33:53.962184 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5z8s4"] Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.062673 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1900-account-create-z9bh8"] Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.440104 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.449930 4831 generic.go:334] "Generic (PLEG): container finished" podID="b639f94c-6740-497d-b53e-4c297f65d750" containerID="013f5a978118240d2d9d2e4e24bccd03590a49ec5061457ba587e7bd2f99cae0" exitCode=0 Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.450034 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8vhg7" event={"ID":"b639f94c-6740-497d-b53e-4c297f65d750","Type":"ContainerDied","Data":"013f5a978118240d2d9d2e4e24bccd03590a49ec5061457ba587e7bd2f99cae0"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.450087 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8vhg7" event={"ID":"b639f94c-6740-497d-b53e-4c297f65d750","Type":"ContainerStarted","Data":"11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.460063 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8db8fec2-ed06-4ad5-8fcf-a491ff18abd1","Type":"ContainerStarted","Data":"9375cf09e1aab8d9226c2c3c1e6f011c069ea7cc159af0febcbe8887666f8c96"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.464737 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8f65-account-create-96zkh" event={"ID":"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe","Type":"ContainerStarted","Data":"3c8ed503732d4ec2e1bec2fe4d9787a75cc99ab96e5c41247be24af5896773a4"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.464793 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8f65-account-create-96zkh" event={"ID":"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe","Type":"ContainerStarted","Data":"fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.474436 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zdfb7" event={"ID":"a6e8cfc9-9958-4540-9b4b-3166796bdbd5","Type":"ContainerStarted","Data":"c438e80f8e8fb25efa46f5b24d875c673edaad273345855a8bcd82af2fa4b5f5"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.474479 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zdfb7" event={"ID":"a6e8cfc9-9958-4540-9b4b-3166796bdbd5","Type":"ContainerStarted","Data":"bdb934c10570853b30699182f2e775371db4ad89e4141b769ad1ea813e1b55d5"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.485446 4831 generic.go:334] "Generic (PLEG): container finished" podID="6aee4d7a-3f47-4df4-ba1b-94af4e39a495" containerID="25d33da447560425d95c099131214a7ee1e8e06c809f30e2ef70a745e517c03b" exitCode=0 Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.485505 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ed31-account-create-fw7hs" event={"ID":"6aee4d7a-3f47-4df4-ba1b-94af4e39a495","Type":"ContainerDied","Data":"25d33da447560425d95c099131214a7ee1e8e06c809f30e2ef70a745e517c03b"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.485529 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ed31-account-create-fw7hs" event={"ID":"6aee4d7a-3f47-4df4-ba1b-94af4e39a495","Type":"ContainerStarted","Data":"28ba23dac24b2ef038a25a26c9efff0bbe2f9964d20033f311198127a78c73e2"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.495683 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z8s4" event={"ID":"7b4819b3-74f6-4015-a437-412e5f4dff2d","Type":"ContainerStarted","Data":"76d923d0c152a7bfda17f33fb79382221ba13750f2fe9b04e6fdf91a4ba9c901"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.495738 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z8s4" event={"ID":"7b4819b3-74f6-4015-a437-412e5f4dff2d","Type":"ContainerStarted","Data":"5622879ff0159021535c38ccf70110e41ea723c84ad8abe6e0b4aa569bc7ef59"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.509207 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1900-account-create-z9bh8" event={"ID":"004ad589-f884-4789-a566-e2076559de28","Type":"ContainerStarted","Data":"45edcf2b7d1862896e4dfb5ab8621241c6f1e13ebd0f8b512a406a8810b79b59"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.509262 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1900-account-create-z9bh8" event={"ID":"004ad589-f884-4789-a566-e2076559de28","Type":"ContainerStarted","Data":"ae82ebc9d3613c9ce5ab560d0b93cf28b5df78d95b4cda0e64b8ac0705959874"} Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.565412 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.626198 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-8f65-account-create-96zkh" podStartSLOduration=8.626180087 podStartE2EDuration="8.626180087s" podCreationTimestamp="2025-11-24 08:33:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:54.58493083 +0000 UTC m=+1108.460075993" watchObservedRunningTime="2025-11-24 08:33:54.626180087 +0000 UTC m=+1108.501325240" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.626768 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.662882 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:33:54 crc kubenswrapper[4831]: E1124 08:33:54.663518 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="sg-core" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.663650 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="sg-core" Nov 24 08:33:54 crc kubenswrapper[4831]: E1124 08:33:54.663732 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="proxy-httpd" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.663785 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="proxy-httpd" Nov 24 08:33:54 crc kubenswrapper[4831]: E1124 08:33:54.663853 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="ceilometer-notification-agent" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.663908 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="ceilometer-notification-agent" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.664141 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="proxy-httpd" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.664230 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="ceilometer-notification-agent" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.664301 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" containerName="sg-core" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.665984 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.672354 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.672629 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.674408 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.689270 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.343980319 podStartE2EDuration="18.689246365s" podCreationTimestamp="2025-11-24 08:33:36 +0000 UTC" firstStartedPulling="2025-11-24 08:33:37.822419699 +0000 UTC m=+1091.697564862" lastFinishedPulling="2025-11-24 08:33:53.167685755 +0000 UTC m=+1107.042830908" observedRunningTime="2025-11-24 08:33:54.614705739 +0000 UTC m=+1108.489850902" watchObservedRunningTime="2025-11-24 08:33:54.689246365 +0000 UTC m=+1108.564391538" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.702358 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1900-account-create-z9bh8" podStartSLOduration=9.702337398 podStartE2EDuration="9.702337398s" podCreationTimestamp="2025-11-24 08:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:54.637702095 +0000 UTC m=+1108.512847248" watchObservedRunningTime="2025-11-24 08:33:54.702337398 +0000 UTC m=+1108.577482551" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.710132 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-5z8s4" podStartSLOduration=9.710092009 podStartE2EDuration="9.710092009s" podCreationTimestamp="2025-11-24 08:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:54.653532967 +0000 UTC m=+1108.528678130" watchObservedRunningTime="2025-11-24 08:33:54.710092009 +0000 UTC m=+1108.585237162" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.733997 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-zdfb7" podStartSLOduration=9.73397383 podStartE2EDuration="9.73397383s" podCreationTimestamp="2025-11-24 08:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:33:54.672010463 +0000 UTC m=+1108.547155616" watchObservedRunningTime="2025-11-24 08:33:54.73397383 +0000 UTC m=+1108.609118983" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760787 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760844 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760861 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760879 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760922 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.760960 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8wr\" (UniqueName: \"kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.761002 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863126 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863179 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863200 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863221 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863265 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863299 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8wr\" (UniqueName: \"kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863362 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.863850 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.865664 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.872564 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.873169 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.873410 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.880563 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.882875 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8wr\" (UniqueName: \"kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr\") pod \"ceilometer-0\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " pod="openstack/ceilometer-0" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.914078 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b" path="/var/lib/kubelet/pods/5b4e7b6e-f5a0-46f4-a2d2-52e96636d21b/volumes" Nov 24 08:33:54 crc kubenswrapper[4831]: I1124 08:33:54.987246 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.451935 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.518591 4831 generic.go:334] "Generic (PLEG): container finished" podID="a6e8cfc9-9958-4540-9b4b-3166796bdbd5" containerID="c438e80f8e8fb25efa46f5b24d875c673edaad273345855a8bcd82af2fa4b5f5" exitCode=0 Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.518661 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zdfb7" event={"ID":"a6e8cfc9-9958-4540-9b4b-3166796bdbd5","Type":"ContainerDied","Data":"c438e80f8e8fb25efa46f5b24d875c673edaad273345855a8bcd82af2fa4b5f5"} Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.520283 4831 generic.go:334] "Generic (PLEG): container finished" podID="7b4819b3-74f6-4015-a437-412e5f4dff2d" containerID="76d923d0c152a7bfda17f33fb79382221ba13750f2fe9b04e6fdf91a4ba9c901" exitCode=0 Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.520431 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z8s4" event={"ID":"7b4819b3-74f6-4015-a437-412e5f4dff2d","Type":"ContainerDied","Data":"76d923d0c152a7bfda17f33fb79382221ba13750f2fe9b04e6fdf91a4ba9c901"} Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.521796 4831 generic.go:334] "Generic (PLEG): container finished" podID="004ad589-f884-4789-a566-e2076559de28" containerID="45edcf2b7d1862896e4dfb5ab8621241c6f1e13ebd0f8b512a406a8810b79b59" exitCode=0 Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.521849 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1900-account-create-z9bh8" event={"ID":"004ad589-f884-4789-a566-e2076559de28","Type":"ContainerDied","Data":"45edcf2b7d1862896e4dfb5ab8621241c6f1e13ebd0f8b512a406a8810b79b59"} Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.523979 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerStarted","Data":"4b5c1ea89f6b6a783ddf901610488422672e68d5676ef6ae34f3b99fb9cbaa82"} Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.529599 4831 generic.go:334] "Generic (PLEG): container finished" podID="2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" containerID="3c8ed503732d4ec2e1bec2fe4d9787a75cc99ab96e5c41247be24af5896773a4" exitCode=0 Nov 24 08:33:55 crc kubenswrapper[4831]: I1124 08:33:55.529704 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8f65-account-create-96zkh" event={"ID":"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe","Type":"ContainerDied","Data":"3c8ed503732d4ec2e1bec2fe4d9787a75cc99ab96e5c41247be24af5896773a4"} Nov 24 08:33:56 crc kubenswrapper[4831]: E1124 08:33:56.177172 4831 kubelet_node_status.go:756] "Failed to set some node status fields" err="failed to validate nodeIP: route ip+net: no such network interface" node="crc" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.455720 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.464908 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.540405 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8vhg7" event={"ID":"b639f94c-6740-497d-b53e-4c297f65d750","Type":"ContainerDied","Data":"11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8"} Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.540447 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11874f32847e74b3b9490c2471ed709a856b3cf43c67eefbdf1655aaf1a544f8" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.540525 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8vhg7" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.545000 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp496\" (UniqueName: \"kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496\") pod \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.545208 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts\") pod \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\" (UID: \"6aee4d7a-3f47-4df4-ba1b-94af4e39a495\") " Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.545571 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ed31-account-create-fw7hs" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.545574 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ed31-account-create-fw7hs" event={"ID":"6aee4d7a-3f47-4df4-ba1b-94af4e39a495","Type":"ContainerDied","Data":"28ba23dac24b2ef038a25a26c9efff0bbe2f9964d20033f311198127a78c73e2"} Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.545613 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28ba23dac24b2ef038a25a26c9efff0bbe2f9964d20033f311198127a78c73e2" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.546387 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6aee4d7a-3f47-4df4-ba1b-94af4e39a495" (UID: "6aee4d7a-3f47-4df4-ba1b-94af4e39a495"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.577988 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496" (OuterVolumeSpecName: "kube-api-access-cp496") pod "6aee4d7a-3f47-4df4-ba1b-94af4e39a495" (UID: "6aee4d7a-3f47-4df4-ba1b-94af4e39a495"). InnerVolumeSpecName "kube-api-access-cp496". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.646828 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt9ln\" (UniqueName: \"kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln\") pod \"b639f94c-6740-497d-b53e-4c297f65d750\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.647331 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts\") pod \"b639f94c-6740-497d-b53e-4c297f65d750\" (UID: \"b639f94c-6740-497d-b53e-4c297f65d750\") " Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.647714 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b639f94c-6740-497d-b53e-4c297f65d750" (UID: "b639f94c-6740-497d-b53e-4c297f65d750"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.648036 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp496\" (UniqueName: \"kubernetes.io/projected/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-kube-api-access-cp496\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.648054 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b639f94c-6740-497d-b53e-4c297f65d750-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.648063 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6aee4d7a-3f47-4df4-ba1b-94af4e39a495-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.651480 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln" (OuterVolumeSpecName: "kube-api-access-zt9ln") pod "b639f94c-6740-497d-b53e-4c297f65d750" (UID: "b639f94c-6740-497d-b53e-4c297f65d750"). InnerVolumeSpecName "kube-api-access-zt9ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.749415 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt9ln\" (UniqueName: \"kubernetes.io/projected/b639f94c-6740-497d-b53e-4c297f65d750-kube-api-access-zt9ln\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:56 crc kubenswrapper[4831]: I1124 08:33:56.936773 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.061404 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts\") pod \"7b4819b3-74f6-4015-a437-412e5f4dff2d\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.061625 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc74r\" (UniqueName: \"kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r\") pod \"7b4819b3-74f6-4015-a437-412e5f4dff2d\" (UID: \"7b4819b3-74f6-4015-a437-412e5f4dff2d\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.063792 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7b4819b3-74f6-4015-a437-412e5f4dff2d" (UID: "7b4819b3-74f6-4015-a437-412e5f4dff2d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.068536 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r" (OuterVolumeSpecName: "kube-api-access-dc74r") pod "7b4819b3-74f6-4015-a437-412e5f4dff2d" (UID: "7b4819b3-74f6-4015-a437-412e5f4dff2d"). InnerVolumeSpecName "kube-api-access-dc74r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.127216 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.140580 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.160417 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.163721 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc74r\" (UniqueName: \"kubernetes.io/projected/7b4819b3-74f6-4015-a437-412e5f4dff2d-kube-api-access-dc74r\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.163746 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b4819b3-74f6-4015-a437-412e5f4dff2d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.266976 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts\") pod \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.267349 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcxfs\" (UniqueName: \"kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs\") pod \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\" (UID: \"a6e8cfc9-9958-4540-9b4b-3166796bdbd5\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.267441 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8lxf\" (UniqueName: \"kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf\") pod \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.267574 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts\") pod \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\" (UID: \"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.267614 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts\") pod \"004ad589-f884-4789-a566-e2076559de28\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.267632 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqfls\" (UniqueName: \"kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls\") pod \"004ad589-f884-4789-a566-e2076559de28\" (UID: \"004ad589-f884-4789-a566-e2076559de28\") " Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.268294 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" (UID: "2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.268719 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "004ad589-f884-4789-a566-e2076559de28" (UID: "004ad589-f884-4789-a566-e2076559de28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.269050 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6e8cfc9-9958-4540-9b4b-3166796bdbd5" (UID: "a6e8cfc9-9958-4540-9b4b-3166796bdbd5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.271821 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls" (OuterVolumeSpecName: "kube-api-access-sqfls") pod "004ad589-f884-4789-a566-e2076559de28" (UID: "004ad589-f884-4789-a566-e2076559de28"). InnerVolumeSpecName "kube-api-access-sqfls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.272985 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf" (OuterVolumeSpecName: "kube-api-access-x8lxf") pod "2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" (UID: "2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe"). InnerVolumeSpecName "kube-api-access-x8lxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.273121 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs" (OuterVolumeSpecName: "kube-api-access-zcxfs") pod "a6e8cfc9-9958-4540-9b4b-3166796bdbd5" (UID: "a6e8cfc9-9958-4540-9b4b-3166796bdbd5"). InnerVolumeSpecName "kube-api-access-zcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369149 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369180 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcxfs\" (UniqueName: \"kubernetes.io/projected/a6e8cfc9-9958-4540-9b4b-3166796bdbd5-kube-api-access-zcxfs\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369193 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8lxf\" (UniqueName: \"kubernetes.io/projected/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-kube-api-access-x8lxf\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369202 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369212 4831 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/004ad589-f884-4789-a566-e2076559de28-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.369221 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqfls\" (UniqueName: \"kubernetes.io/projected/004ad589-f884-4789-a566-e2076559de28-kube-api-access-sqfls\") on node \"crc\" DevicePath \"\"" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.556372 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-8f65-account-create-96zkh" event={"ID":"2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe","Type":"ContainerDied","Data":"fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea"} Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.556414 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcc053d9bf4809dd8ec89ce4c1d99133ef358ef46b452b9ed3fdd7fcfc5d38ea" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.556498 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-8f65-account-create-96zkh" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.569062 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zdfb7" event={"ID":"a6e8cfc9-9958-4540-9b4b-3166796bdbd5","Type":"ContainerDied","Data":"bdb934c10570853b30699182f2e775371db4ad89e4141b769ad1ea813e1b55d5"} Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.569097 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdb934c10570853b30699182f2e775371db4ad89e4141b769ad1ea813e1b55d5" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.569147 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zdfb7" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.574623 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5z8s4" event={"ID":"7b4819b3-74f6-4015-a437-412e5f4dff2d","Type":"ContainerDied","Data":"5622879ff0159021535c38ccf70110e41ea723c84ad8abe6e0b4aa569bc7ef59"} Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.574655 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5622879ff0159021535c38ccf70110e41ea723c84ad8abe6e0b4aa569bc7ef59" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.574702 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5z8s4" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.578257 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1900-account-create-z9bh8" event={"ID":"004ad589-f884-4789-a566-e2076559de28","Type":"ContainerDied","Data":"ae82ebc9d3613c9ce5ab560d0b93cf28b5df78d95b4cda0e64b8ac0705959874"} Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.578453 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae82ebc9d3613c9ce5ab560d0b93cf28b5df78d95b4cda0e64b8ac0705959874" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.578594 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1900-account-create-z9bh8" Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.590651 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerStarted","Data":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} Nov 24 08:33:57 crc kubenswrapper[4831]: I1124 08:33:57.590899 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerStarted","Data":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} Nov 24 08:33:58 crc kubenswrapper[4831]: I1124 08:33:58.401804 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:33:58 crc kubenswrapper[4831]: I1124 08:33:58.402264 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:33:58 crc kubenswrapper[4831]: I1124 08:33:58.600305 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerStarted","Data":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} Nov 24 08:34:00 crc kubenswrapper[4831]: I1124 08:34:00.629697 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerStarted","Data":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} Nov 24 08:34:00 crc kubenswrapper[4831]: I1124 08:34:00.631646 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:34:00 crc kubenswrapper[4831]: I1124 08:34:00.664905 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.735226478 podStartE2EDuration="6.664879066s" podCreationTimestamp="2025-11-24 08:33:54 +0000 UTC" firstStartedPulling="2025-11-24 08:33:55.462233078 +0000 UTC m=+1109.337378231" lastFinishedPulling="2025-11-24 08:33:59.391885666 +0000 UTC m=+1113.267030819" observedRunningTime="2025-11-24 08:34:00.659839682 +0000 UTC m=+1114.534984875" watchObservedRunningTime="2025-11-24 08:34:00.664879066 +0000 UTC m=+1114.540024229" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344014 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ckzq6"] Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344457 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b4819b3-74f6-4015-a437-412e5f4dff2d" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344480 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b4819b3-74f6-4015-a437-412e5f4dff2d" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344504 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344512 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344526 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aee4d7a-3f47-4df4-ba1b-94af4e39a495" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344533 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aee4d7a-3f47-4df4-ba1b-94af4e39a495" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344544 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6e8cfc9-9958-4540-9b4b-3166796bdbd5" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344550 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6e8cfc9-9958-4540-9b4b-3166796bdbd5" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344564 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b639f94c-6740-497d-b53e-4c297f65d750" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344572 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b639f94c-6740-497d-b53e-4c297f65d750" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: E1124 08:34:01.344595 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="004ad589-f884-4789-a566-e2076559de28" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344602 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="004ad589-f884-4789-a566-e2076559de28" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344753 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b4819b3-74f6-4015-a437-412e5f4dff2d" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344768 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344782 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="004ad589-f884-4789-a566-e2076559de28" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344793 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6e8cfc9-9958-4540-9b4b-3166796bdbd5" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344803 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aee4d7a-3f47-4df4-ba1b-94af4e39a495" containerName="mariadb-account-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.344815 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b639f94c-6740-497d-b53e-4c297f65d750" containerName="mariadb-database-create" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.345436 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.349429 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.349531 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.349664 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t5mqt" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.408819 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ckzq6"] Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.447857 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.447992 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.448026 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.448077 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8pqq\" (UniqueName: \"kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.550061 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.550178 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.550224 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.550279 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8pqq\" (UniqueName: \"kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.566203 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.568579 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8pqq\" (UniqueName: \"kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.569160 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.569248 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data\") pod \"nova-cell0-conductor-db-sync-ckzq6\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:01 crc kubenswrapper[4831]: I1124 08:34:01.662236 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:02 crc kubenswrapper[4831]: I1124 08:34:02.218226 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ckzq6"] Nov 24 08:34:02 crc kubenswrapper[4831]: I1124 08:34:02.642852 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" event={"ID":"ff4e0931-319b-4ee2-bbef-224d6393f94a","Type":"ContainerStarted","Data":"b3d396efdaa3e555ef21d42df5e7ff73aacddf694bb3349da04f2aaf538e4e72"} Nov 24 08:34:04 crc kubenswrapper[4831]: I1124 08:34:04.568211 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:04 crc kubenswrapper[4831]: I1124 08:34:04.568720 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-central-agent" containerID="cri-o://0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" gracePeriod=30 Nov 24 08:34:04 crc kubenswrapper[4831]: I1124 08:34:04.569120 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="sg-core" containerID="cri-o://50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" gracePeriod=30 Nov 24 08:34:04 crc kubenswrapper[4831]: I1124 08:34:04.569164 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="proxy-httpd" containerID="cri-o://1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" gracePeriod=30 Nov 24 08:34:04 crc kubenswrapper[4831]: I1124 08:34:04.569208 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-notification-agent" containerID="cri-o://19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" gracePeriod=30 Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.446500 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620506 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620580 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620683 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620767 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm8wr\" (UniqueName: \"kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620806 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620822 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.620870 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts\") pod \"1d416cc1-e545-4f28-9dee-56b03eda6b35\" (UID: \"1d416cc1-e545-4f28-9dee-56b03eda6b35\") " Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.622444 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.622604 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.631153 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr" (OuterVolumeSpecName: "kube-api-access-gm8wr") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "kube-api-access-gm8wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.631663 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts" (OuterVolumeSpecName: "scripts") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.651015 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683237 4831 generic.go:334] "Generic (PLEG): container finished" podID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" exitCode=0 Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683267 4831 generic.go:334] "Generic (PLEG): container finished" podID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" exitCode=2 Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683275 4831 generic.go:334] "Generic (PLEG): container finished" podID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" exitCode=0 Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683283 4831 generic.go:334] "Generic (PLEG): container finished" podID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" exitCode=0 Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683305 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerDied","Data":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683342 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerDied","Data":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683353 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerDied","Data":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683362 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerDied","Data":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683373 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1d416cc1-e545-4f28-9dee-56b03eda6b35","Type":"ContainerDied","Data":"4b5c1ea89f6b6a783ddf901610488422672e68d5676ef6ae34f3b99fb9cbaa82"} Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683388 4831 scope.go:117] "RemoveContainer" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.683411 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.733866 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734938 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm8wr\" (UniqueName: \"kubernetes.io/projected/1d416cc1-e545-4f28-9dee-56b03eda6b35-kube-api-access-gm8wr\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734956 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734965 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734974 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734985 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.734997 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1d416cc1-e545-4f28-9dee-56b03eda6b35-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.764066 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data" (OuterVolumeSpecName: "config-data") pod "1d416cc1-e545-4f28-9dee-56b03eda6b35" (UID: "1d416cc1-e545-4f28-9dee-56b03eda6b35"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.819626 4831 scope.go:117] "RemoveContainer" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.836880 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d416cc1-e545-4f28-9dee-56b03eda6b35-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.840106 4831 scope.go:117] "RemoveContainer" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.856691 4831 scope.go:117] "RemoveContainer" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.879791 4831 scope.go:117] "RemoveContainer" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: E1124 08:34:05.880402 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": container with ID starting with 1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c not found: ID does not exist" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.880441 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} err="failed to get container status \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": rpc error: code = NotFound desc = could not find container \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": container with ID starting with 1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.880462 4831 scope.go:117] "RemoveContainer" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: E1124 08:34:05.880919 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": container with ID starting with 50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2 not found: ID does not exist" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.880949 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} err="failed to get container status \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": rpc error: code = NotFound desc = could not find container \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": container with ID starting with 50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.880962 4831 scope.go:117] "RemoveContainer" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: E1124 08:34:05.881260 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": container with ID starting with 19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44 not found: ID does not exist" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881287 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} err="failed to get container status \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": rpc error: code = NotFound desc = could not find container \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": container with ID starting with 19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881301 4831 scope.go:117] "RemoveContainer" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: E1124 08:34:05.881613 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": container with ID starting with 0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696 not found: ID does not exist" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881634 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} err="failed to get container status \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": rpc error: code = NotFound desc = could not find container \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": container with ID starting with 0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881647 4831 scope.go:117] "RemoveContainer" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881875 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} err="failed to get container status \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": rpc error: code = NotFound desc = could not find container \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": container with ID starting with 1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.881892 4831 scope.go:117] "RemoveContainer" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882084 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} err="failed to get container status \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": rpc error: code = NotFound desc = could not find container \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": container with ID starting with 50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882099 4831 scope.go:117] "RemoveContainer" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882301 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} err="failed to get container status \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": rpc error: code = NotFound desc = could not find container \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": container with ID starting with 19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882334 4831 scope.go:117] "RemoveContainer" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882610 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} err="failed to get container status \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": rpc error: code = NotFound desc = could not find container \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": container with ID starting with 0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882626 4831 scope.go:117] "RemoveContainer" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882848 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} err="failed to get container status \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": rpc error: code = NotFound desc = could not find container \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": container with ID starting with 1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.882864 4831 scope.go:117] "RemoveContainer" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883057 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} err="failed to get container status \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": rpc error: code = NotFound desc = could not find container \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": container with ID starting with 50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883073 4831 scope.go:117] "RemoveContainer" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883272 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} err="failed to get container status \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": rpc error: code = NotFound desc = could not find container \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": container with ID starting with 19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883289 4831 scope.go:117] "RemoveContainer" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883510 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} err="failed to get container status \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": rpc error: code = NotFound desc = could not find container \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": container with ID starting with 0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883526 4831 scope.go:117] "RemoveContainer" containerID="1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883721 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c"} err="failed to get container status \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": rpc error: code = NotFound desc = could not find container \"1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c\": container with ID starting with 1be05b00ed4a1454c8162381100c189203c3ea58229aadc5a97a06c6a1dde06c not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883736 4831 scope.go:117] "RemoveContainer" containerID="50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.883976 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2"} err="failed to get container status \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": rpc error: code = NotFound desc = could not find container \"50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2\": container with ID starting with 50a0c33ba504635c93a5b714bdaf31397c7ab73864063139dc1d8d8f3a0012f2 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.884011 4831 scope.go:117] "RemoveContainer" containerID="19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.884241 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44"} err="failed to get container status \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": rpc error: code = NotFound desc = could not find container \"19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44\": container with ID starting with 19b1ec1c9926d25a7f62b8ede848dabcc05b775459a8396707cc94c29ff18e44 not found: ID does not exist" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.884258 4831 scope.go:117] "RemoveContainer" containerID="0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696" Nov 24 08:34:05 crc kubenswrapper[4831]: I1124 08:34:05.884568 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696"} err="failed to get container status \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": rpc error: code = NotFound desc = could not find container \"0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696\": container with ID starting with 0819823ef8c946b30763dd24d346a9c9eac2b0004fc7ae73a3b3249716475696 not found: ID does not exist" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.031009 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.049222 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.063902 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:06 crc kubenswrapper[4831]: E1124 08:34:06.064353 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-central-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064375 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-central-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: E1124 08:34:06.064390 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="proxy-httpd" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064398 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="proxy-httpd" Nov 24 08:34:06 crc kubenswrapper[4831]: E1124 08:34:06.064432 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="sg-core" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064439 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="sg-core" Nov 24 08:34:06 crc kubenswrapper[4831]: E1124 08:34:06.064456 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-notification-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064463 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-notification-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064648 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="sg-core" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064664 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-central-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064674 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="ceilometer-notification-agent" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.064690 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" containerName="proxy-httpd" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.066076 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.066165 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.076021 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.076219 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243256 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243455 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243540 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8xks\" (UniqueName: \"kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243601 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243706 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243914 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.243988 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345462 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345571 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345602 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8xks\" (UniqueName: \"kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345649 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345681 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345769 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.345843 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.346480 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.353423 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.353862 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.353975 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.363155 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.363888 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8xks\" (UniqueName: \"kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.364092 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.396522 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:06 crc kubenswrapper[4831]: I1124 08:34:06.919213 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d416cc1-e545-4f28-9dee-56b03eda6b35" path="/var/lib/kubelet/pods/1d416cc1-e545-4f28-9dee-56b03eda6b35/volumes" Nov 24 08:34:11 crc kubenswrapper[4831]: I1124 08:34:11.234270 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:11 crc kubenswrapper[4831]: W1124 08:34:11.234757 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cced89f_0754_4e67_83b4_784b8cbd435e.slice/crio-094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655 WatchSource:0}: Error finding container 094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655: Status 404 returned error can't find the container with id 094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655 Nov 24 08:34:11 crc kubenswrapper[4831]: I1124 08:34:11.746270 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" event={"ID":"ff4e0931-319b-4ee2-bbef-224d6393f94a","Type":"ContainerStarted","Data":"e64e5de0eb9f0070417fe8b2a65d406a7511d5e8d42016b1916a01e64cff2573"} Nov 24 08:34:11 crc kubenswrapper[4831]: I1124 08:34:11.747535 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerStarted","Data":"094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655"} Nov 24 08:34:11 crc kubenswrapper[4831]: I1124 08:34:11.770493 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" podStartSLOduration=2.171333222 podStartE2EDuration="10.770474044s" podCreationTimestamp="2025-11-24 08:34:01 +0000 UTC" firstStartedPulling="2025-11-24 08:34:02.262575817 +0000 UTC m=+1116.137720970" lastFinishedPulling="2025-11-24 08:34:10.861716629 +0000 UTC m=+1124.736861792" observedRunningTime="2025-11-24 08:34:11.762804265 +0000 UTC m=+1125.637949438" watchObservedRunningTime="2025-11-24 08:34:11.770474044 +0000 UTC m=+1125.645619207" Nov 24 08:34:12 crc kubenswrapper[4831]: I1124 08:34:12.757730 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerStarted","Data":"e39f033bcc7de8d23a03832a19c362479924e6568ace2b408c2151f5a8a816ac"} Nov 24 08:34:12 crc kubenswrapper[4831]: I1124 08:34:12.758366 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerStarted","Data":"4a5ed1fb1f17b795132bc60e9852405f12f91249a8ecb2e6776d61061ee5b25b"} Nov 24 08:34:14 crc kubenswrapper[4831]: I1124 08:34:14.514448 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:14 crc kubenswrapper[4831]: I1124 08:34:14.778054 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerStarted","Data":"38ac015fbf7f9512a409b657f175e1e635b57d0593706072dd06e01f47e139b9"} Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.788070 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerStarted","Data":"058b4381d164c1e31d21195c601da866731483fa997f882ddd74acce5522639b"} Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.788704 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-central-agent" containerID="cri-o://4a5ed1fb1f17b795132bc60e9852405f12f91249a8ecb2e6776d61061ee5b25b" gracePeriod=30 Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.788760 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="proxy-httpd" containerID="cri-o://058b4381d164c1e31d21195c601da866731483fa997f882ddd74acce5522639b" gracePeriod=30 Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.788798 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="sg-core" containerID="cri-o://38ac015fbf7f9512a409b657f175e1e635b57d0593706072dd06e01f47e139b9" gracePeriod=30 Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.788838 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-notification-agent" containerID="cri-o://e39f033bcc7de8d23a03832a19c362479924e6568ace2b408c2151f5a8a816ac" gracePeriod=30 Nov 24 08:34:15 crc kubenswrapper[4831]: I1124 08:34:15.789126 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801167 4831 generic.go:334] "Generic (PLEG): container finished" podID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerID="058b4381d164c1e31d21195c601da866731483fa997f882ddd74acce5522639b" exitCode=0 Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801696 4831 generic.go:334] "Generic (PLEG): container finished" podID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerID="38ac015fbf7f9512a409b657f175e1e635b57d0593706072dd06e01f47e139b9" exitCode=2 Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801717 4831 generic.go:334] "Generic (PLEG): container finished" podID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerID="e39f033bcc7de8d23a03832a19c362479924e6568ace2b408c2151f5a8a816ac" exitCode=0 Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801195 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerDied","Data":"058b4381d164c1e31d21195c601da866731483fa997f882ddd74acce5522639b"} Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801777 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerDied","Data":"38ac015fbf7f9512a409b657f175e1e635b57d0593706072dd06e01f47e139b9"} Nov 24 08:34:16 crc kubenswrapper[4831]: I1124 08:34:16.801802 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerDied","Data":"e39f033bcc7de8d23a03832a19c362479924e6568ace2b408c2151f5a8a816ac"} Nov 24 08:34:20 crc kubenswrapper[4831]: I1124 08:34:20.838077 4831 generic.go:334] "Generic (PLEG): container finished" podID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerID="4a5ed1fb1f17b795132bc60e9852405f12f91249a8ecb2e6776d61061ee5b25b" exitCode=0 Nov 24 08:34:20 crc kubenswrapper[4831]: I1124 08:34:20.838140 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerDied","Data":"4a5ed1fb1f17b795132bc60e9852405f12f91249a8ecb2e6776d61061ee5b25b"} Nov 24 08:34:20 crc kubenswrapper[4831]: I1124 08:34:20.838745 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cced89f-0754-4e67-83b4-784b8cbd435e","Type":"ContainerDied","Data":"094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655"} Nov 24 08:34:20 crc kubenswrapper[4831]: I1124 08:34:20.838810 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="094956e0f2f1909d4b5aa7750afa7eeecc837f752e5e59ae86b482393999c655" Nov 24 08:34:20 crc kubenswrapper[4831]: I1124 08:34:20.880067 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.022567 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.022702 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.022886 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.023055 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8xks\" (UniqueName: \"kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.023096 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.023127 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.023175 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd\") pod \"0cced89f-0754-4e67-83b4-784b8cbd435e\" (UID: \"0cced89f-0754-4e67-83b4-784b8cbd435e\") " Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.024557 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.024604 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.031832 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks" (OuterVolumeSpecName: "kube-api-access-r8xks") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "kube-api-access-r8xks". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.032412 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts" (OuterVolumeSpecName: "scripts") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.046941 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.106939 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126227 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126255 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126266 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8xks\" (UniqueName: \"kubernetes.io/projected/0cced89f-0754-4e67-83b4-784b8cbd435e-kube-api-access-r8xks\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126278 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126289 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cced89f-0754-4e67-83b4-784b8cbd435e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.126298 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.127374 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data" (OuterVolumeSpecName: "config-data") pod "0cced89f-0754-4e67-83b4-784b8cbd435e" (UID: "0cced89f-0754-4e67-83b4-784b8cbd435e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.228552 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cced89f-0754-4e67-83b4-784b8cbd435e-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.848377 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.885854 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.893858 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917092 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:21 crc kubenswrapper[4831]: E1124 08:34:21.917652 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-notification-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917677 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-notification-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: E1124 08:34:21.917704 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-central-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917713 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-central-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: E1124 08:34:21.917731 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="proxy-httpd" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917740 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="proxy-httpd" Nov 24 08:34:21 crc kubenswrapper[4831]: E1124 08:34:21.917767 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="sg-core" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917774 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="sg-core" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917953 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-notification-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917967 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="proxy-httpd" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917983 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="ceilometer-central-agent" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.917995 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" containerName="sg-core" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.919453 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.923550 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.924408 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.927416 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948465 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p55bm\" (UniqueName: \"kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948520 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948562 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948595 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948699 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948764 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:21 crc kubenswrapper[4831]: I1124 08:34:21.948777 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.049961 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p55bm\" (UniqueName: \"kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050014 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050049 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050076 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050113 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050159 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.050172 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.051407 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.051532 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.056520 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.057351 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.058277 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.062995 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.080570 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p55bm\" (UniqueName: \"kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm\") pod \"ceilometer-0\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.248471 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.765408 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.885228 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerStarted","Data":"e53c1c5f14f7137a08016d2cfa0db28f8bb70b00d2e422dc863198f1272c1293"} Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.888079 4831 generic.go:334] "Generic (PLEG): container finished" podID="ff4e0931-319b-4ee2-bbef-224d6393f94a" containerID="e64e5de0eb9f0070417fe8b2a65d406a7511d5e8d42016b1916a01e64cff2573" exitCode=0 Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.888252 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" event={"ID":"ff4e0931-319b-4ee2-bbef-224d6393f94a","Type":"ContainerDied","Data":"e64e5de0eb9f0070417fe8b2a65d406a7511d5e8d42016b1916a01e64cff2573"} Nov 24 08:34:22 crc kubenswrapper[4831]: I1124 08:34:22.913630 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cced89f-0754-4e67-83b4-784b8cbd435e" path="/var/lib/kubelet/pods/0cced89f-0754-4e67-83b4-784b8cbd435e/volumes" Nov 24 08:34:23 crc kubenswrapper[4831]: I1124 08:34:23.900623 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerStarted","Data":"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793"} Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.224640 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.292209 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data\") pod \"ff4e0931-319b-4ee2-bbef-224d6393f94a\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.292944 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8pqq\" (UniqueName: \"kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq\") pod \"ff4e0931-319b-4ee2-bbef-224d6393f94a\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.293208 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts\") pod \"ff4e0931-319b-4ee2-bbef-224d6393f94a\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.293369 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle\") pod \"ff4e0931-319b-4ee2-bbef-224d6393f94a\" (UID: \"ff4e0931-319b-4ee2-bbef-224d6393f94a\") " Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.299501 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts" (OuterVolumeSpecName: "scripts") pod "ff4e0931-319b-4ee2-bbef-224d6393f94a" (UID: "ff4e0931-319b-4ee2-bbef-224d6393f94a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.299783 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq" (OuterVolumeSpecName: "kube-api-access-w8pqq") pod "ff4e0931-319b-4ee2-bbef-224d6393f94a" (UID: "ff4e0931-319b-4ee2-bbef-224d6393f94a"). InnerVolumeSpecName "kube-api-access-w8pqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.316409 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data" (OuterVolumeSpecName: "config-data") pod "ff4e0931-319b-4ee2-bbef-224d6393f94a" (UID: "ff4e0931-319b-4ee2-bbef-224d6393f94a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.327150 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff4e0931-319b-4ee2-bbef-224d6393f94a" (UID: "ff4e0931-319b-4ee2-bbef-224d6393f94a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.396831 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.396879 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.396894 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff4e0931-319b-4ee2-bbef-224d6393f94a-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.396906 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8pqq\" (UniqueName: \"kubernetes.io/projected/ff4e0931-319b-4ee2-bbef-224d6393f94a-kube-api-access-w8pqq\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.915393 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" event={"ID":"ff4e0931-319b-4ee2-bbef-224d6393f94a","Type":"ContainerDied","Data":"b3d396efdaa3e555ef21d42df5e7ff73aacddf694bb3349da04f2aaf538e4e72"} Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.915777 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3d396efdaa3e555ef21d42df5e7ff73aacddf694bb3349da04f2aaf538e4e72" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.915846 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ckzq6" Nov 24 08:34:24 crc kubenswrapper[4831]: I1124 08:34:24.941637 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerStarted","Data":"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35"} Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.008802 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:34:25 crc kubenswrapper[4831]: E1124 08:34:25.009469 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff4e0931-319b-4ee2-bbef-224d6393f94a" containerName="nova-cell0-conductor-db-sync" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.009496 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff4e0931-319b-4ee2-bbef-224d6393f94a" containerName="nova-cell0-conductor-db-sync" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.009696 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff4e0931-319b-4ee2-bbef-224d6393f94a" containerName="nova-cell0-conductor-db-sync" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.010516 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.013068 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.013936 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t5mqt" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.023749 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.117096 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.117289 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwt7m\" (UniqueName: \"kubernetes.io/projected/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-kube-api-access-wwt7m\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.117361 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.219279 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwt7m\" (UniqueName: \"kubernetes.io/projected/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-kube-api-access-wwt7m\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.219441 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.219493 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.234621 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.235139 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.239838 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwt7m\" (UniqueName: \"kubernetes.io/projected/d798294b-fc7d-4ec4-8ad6-5bf73c68144d-kube-api-access-wwt7m\") pod \"nova-cell0-conductor-0\" (UID: \"d798294b-fc7d-4ec4-8ad6-5bf73c68144d\") " pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.340772 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.789163 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 24 08:34:25 crc kubenswrapper[4831]: W1124 08:34:25.795610 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd798294b_fc7d_4ec4_8ad6_5bf73c68144d.slice/crio-4b8ff9b244434e5b5a9fb93ad3b92eee4337fc222fa55504a221821c00d36a4e WatchSource:0}: Error finding container 4b8ff9b244434e5b5a9fb93ad3b92eee4337fc222fa55504a221821c00d36a4e: Status 404 returned error can't find the container with id 4b8ff9b244434e5b5a9fb93ad3b92eee4337fc222fa55504a221821c00d36a4e Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.951548 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d798294b-fc7d-4ec4-8ad6-5bf73c68144d","Type":"ContainerStarted","Data":"4b8ff9b244434e5b5a9fb93ad3b92eee4337fc222fa55504a221821c00d36a4e"} Nov 24 08:34:25 crc kubenswrapper[4831]: I1124 08:34:25.954071 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerStarted","Data":"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb"} Nov 24 08:34:26 crc kubenswrapper[4831]: I1124 08:34:26.970723 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d798294b-fc7d-4ec4-8ad6-5bf73c68144d","Type":"ContainerStarted","Data":"41e070dfe18fb39bb5fc69630ccd5d75c470ab8ce3d2a2b46bc07c97cbdce0bf"} Nov 24 08:34:26 crc kubenswrapper[4831]: I1124 08:34:26.971433 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:26 crc kubenswrapper[4831]: I1124 08:34:26.977631 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerStarted","Data":"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d"} Nov 24 08:34:26 crc kubenswrapper[4831]: I1124 08:34:26.978404 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:34:26 crc kubenswrapper[4831]: I1124 08:34:26.997810 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.9977880470000002 podStartE2EDuration="2.997788047s" podCreationTimestamp="2025-11-24 08:34:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:26.986020512 +0000 UTC m=+1140.861165675" watchObservedRunningTime="2025-11-24 08:34:26.997788047 +0000 UTC m=+1140.872933210" Nov 24 08:34:27 crc kubenswrapper[4831]: I1124 08:34:27.042889 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.567956732 podStartE2EDuration="6.042871203s" podCreationTimestamp="2025-11-24 08:34:21 +0000 UTC" firstStartedPulling="2025-11-24 08:34:22.796484742 +0000 UTC m=+1136.671629895" lastFinishedPulling="2025-11-24 08:34:26.271399213 +0000 UTC m=+1140.146544366" observedRunningTime="2025-11-24 08:34:27.032739114 +0000 UTC m=+1140.907884277" watchObservedRunningTime="2025-11-24 08:34:27.042871203 +0000 UTC m=+1140.918016356" Nov 24 08:34:28 crc kubenswrapper[4831]: I1124 08:34:28.401358 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:34:28 crc kubenswrapper[4831]: I1124 08:34:28.401675 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:34:28 crc kubenswrapper[4831]: I1124 08:34:28.401726 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:34:28 crc kubenswrapper[4831]: I1124 08:34:28.402653 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:34:28 crc kubenswrapper[4831]: I1124 08:34:28.402739 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899" gracePeriod=600 Nov 24 08:34:29 crc kubenswrapper[4831]: I1124 08:34:29.028981 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899" exitCode=0 Nov 24 08:34:29 crc kubenswrapper[4831]: I1124 08:34:29.030402 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899"} Nov 24 08:34:29 crc kubenswrapper[4831]: I1124 08:34:29.030512 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e"} Nov 24 08:34:29 crc kubenswrapper[4831]: I1124 08:34:29.030584 4831 scope.go:117] "RemoveContainer" containerID="8b02d13f96e1930fd8c06fcca913efdd24ced68d43d410f3a05d24217dd53a40" Nov 24 08:34:30 crc kubenswrapper[4831]: I1124 08:34:30.418738 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.059197 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lhjsw"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.060834 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.064384 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.077363 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhjsw"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.083608 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.154203 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pzbz\" (UniqueName: \"kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.154352 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.154381 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.154407 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.255897 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.257183 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.257231 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.257915 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pzbz\" (UniqueName: \"kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.267080 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.275989 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.282249 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.296384 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.297865 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.315422 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.327672 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pzbz\" (UniqueName: \"kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz\") pod \"nova-cell0-cell-mapping-lhjsw\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.347450 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.360643 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgfw5\" (UniqueName: \"kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.360697 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.360822 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.388999 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.463660 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.463742 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgfw5\" (UniqueName: \"kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.463781 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.470738 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.486252 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.491036 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.492816 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.495754 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.512924 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.519700 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgfw5\" (UniqueName: \"kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5\") pod \"nova-scheduler-0\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.549642 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.551360 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.556421 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.564938 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.564979 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjh54\" (UniqueName: \"kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.565006 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.565036 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.580456 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.582909 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.585549 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.639826 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.668050 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669792 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669822 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjh54\" (UniqueName: \"kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669857 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669887 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669947 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.669975 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.670010 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lhw5\" (UniqueName: \"kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.670050 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.670078 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.670108 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c468q\" (UniqueName: \"kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.670166 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.675716 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.692925 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.693852 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.720694 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjh54\" (UniqueName: \"kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54\") pod \"nova-api-0\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771277 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c468q\" (UniqueName: \"kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771366 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771466 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771492 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771516 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lhw5\" (UniqueName: \"kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771553 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.771569 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.772061 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.778060 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.778290 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.779666 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.782142 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.896099 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.915969 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.918891 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lhw5\" (UniqueName: \"kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5\") pod \"nova-cell1-novncproxy-0\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.938352 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.938560 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.946880 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.951049 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c468q\" (UniqueName: \"kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q\") pod \"nova-metadata-0\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " pod="openstack/nova-metadata-0" Nov 24 08:34:31 crc kubenswrapper[4831]: I1124 08:34:31.963702 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.131601 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz7ws\" (UniqueName: \"kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.131873 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.131910 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.133171 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.133220 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.212164 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.238039 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.238121 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.238147 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.238235 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz7ws\" (UniqueName: \"kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.238255 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.239183 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.239690 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.240205 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.240753 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.349506 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz7ws\" (UniqueName: \"kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws\") pod \"dnsmasq-dns-8b8cf6657-26cdj\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.492493 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:32 crc kubenswrapper[4831]: W1124 08:34:32.493560 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd246dd0b_f6b4_4fd0_a4b3_19e830cb38a1.slice/crio-5964d8ee80bbeaccedac973ec82d23117b8bdd387308e64003000be2e75ddb56 WatchSource:0}: Error finding container 5964d8ee80bbeaccedac973ec82d23117b8bdd387308e64003000be2e75ddb56: Status 404 returned error can't find the container with id 5964d8ee80bbeaccedac973ec82d23117b8bdd387308e64003000be2e75ddb56 Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.524251 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhjsw"] Nov 24 08:34:32 crc kubenswrapper[4831]: I1124 08:34:32.608465 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:32.998368 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.027462 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.075095 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.155229 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.159421 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1","Type":"ContainerStarted","Data":"5964d8ee80bbeaccedac973ec82d23117b8bdd387308e64003000be2e75ddb56"} Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.160552 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9d3bf305-0549-4809-acc9-3ea794deec4c","Type":"ContainerStarted","Data":"a970a13b929df4590e2f3ef618fb1eb6049f932948d57aa3baaba3da1a7d17ae"} Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.161609 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerStarted","Data":"0bb563fe29b8dcd54175e1f192cfe8432f47dabe14b86ce91f78f88b4af5e364"} Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.162371 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerStarted","Data":"c13bac6e70f7fdcce1c2938098b3e56825d5511fccf3aca34bf9268b75b2a2b6"} Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.172191 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhjsw" event={"ID":"8fbc0462-5e34-49ff-8786-5806fbbf36ec","Type":"ContainerStarted","Data":"928a9696942d6a2d6dba4c02b1e10fef0fbff64ae29c4bef1d9ee8489547c45c"} Nov 24 08:34:33 crc kubenswrapper[4831]: W1124 08:34:33.176333 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5022f8f0_893d_4e8a_8fb8_b9967ed943d8.slice/crio-b0ecb45204872b7203e1fc16962a5e597f2e193aed01cc58c543e4eac5daf2ba WatchSource:0}: Error finding container b0ecb45204872b7203e1fc16962a5e597f2e193aed01cc58c543e4eac5daf2ba: Status 404 returned error can't find the container with id b0ecb45204872b7203e1fc16962a5e597f2e193aed01cc58c543e4eac5daf2ba Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.801772 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vl2xc"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.803391 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.823695 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.823889 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.874046 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vl2xc"] Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.891716 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.894078 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7c89\" (UniqueName: \"kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.894149 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.894193 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.996345 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.996640 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7c89\" (UniqueName: \"kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.996779 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:33 crc kubenswrapper[4831]: I1124 08:34:33.996869 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.023002 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.025105 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.026059 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7c89\" (UniqueName: \"kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.050202 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data\") pod \"nova-cell1-conductor-db-sync-vl2xc\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.134734 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.217310 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhjsw" event={"ID":"8fbc0462-5e34-49ff-8786-5806fbbf36ec","Type":"ContainerStarted","Data":"e578b78210830b2e0a840fed543cf9da8fcdbf4715c4a33306b37635b129c57f"} Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.220237 4831 generic.go:334] "Generic (PLEG): container finished" podID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerID="a38711d3e958f5678dbac631e43d5558b0c423f716aadc01030c8635dee9fa49" exitCode=0 Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.220273 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" event={"ID":"5022f8f0-893d-4e8a-8fb8-b9967ed943d8","Type":"ContainerDied","Data":"a38711d3e958f5678dbac631e43d5558b0c423f716aadc01030c8635dee9fa49"} Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.220292 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" event={"ID":"5022f8f0-893d-4e8a-8fb8-b9967ed943d8","Type":"ContainerStarted","Data":"b0ecb45204872b7203e1fc16962a5e597f2e193aed01cc58c543e4eac5daf2ba"} Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.252081 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lhjsw" podStartSLOduration=3.25205175 podStartE2EDuration="3.25205175s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:34.240799969 +0000 UTC m=+1148.115945132" watchObservedRunningTime="2025-11-24 08:34:34.25205175 +0000 UTC m=+1148.127196903" Nov 24 08:34:34 crc kubenswrapper[4831]: I1124 08:34:34.583976 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vl2xc"] Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.232165 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" event={"ID":"5022f8f0-893d-4e8a-8fb8-b9967ed943d8","Type":"ContainerStarted","Data":"d9403092dffd26cc22c5dbdc758b999a3397debbf39689e5349ec3428fa6fd6c"} Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.232266 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.235190 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" event={"ID":"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd","Type":"ContainerStarted","Data":"3445fbc41cd91e936065f75aac0f1c1d9680a4191369f85b9b9112ad3670f3b6"} Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.269069 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" podStartSLOduration=4.269047251 podStartE2EDuration="4.269047251s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:35.266718935 +0000 UTC m=+1149.141864108" watchObservedRunningTime="2025-11-24 08:34:35.269047251 +0000 UTC m=+1149.144192404" Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.549850 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:34:35 crc kubenswrapper[4831]: I1124 08:34:35.562021 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:34:37 crc kubenswrapper[4831]: I1124 08:34:37.266021 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" event={"ID":"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd","Type":"ContainerStarted","Data":"847c5e0f764c84d0674b9661873594c47402a7b00fbe858162979eae881f5ac9"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.284494 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1","Type":"ContainerStarted","Data":"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.287231 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9d3bf305-0549-4809-acc9-3ea794deec4c","Type":"ContainerStarted","Data":"c64e6fd12cf514241c834e3d9f449c2e4758b32856e47c5d2b4db7566f5c0cdd"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.287352 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9d3bf305-0549-4809-acc9-3ea794deec4c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c64e6fd12cf514241c834e3d9f449c2e4758b32856e47c5d2b4db7566f5c0cdd" gracePeriod=30 Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.294347 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerStarted","Data":"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.294439 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerStarted","Data":"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.298049 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerStarted","Data":"4883f184578c131b7aea2b549436953ef6a4a1eb55f1d9a1053ca88e7faf8bfd"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.298103 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerStarted","Data":"6ae768aa3b739a7e4daa33dc68d529f0a22bd59d873d4021bd75921fffd7ba5b"} Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.298158 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-log" containerID="cri-o://6ae768aa3b739a7e4daa33dc68d529f0a22bd59d873d4021bd75921fffd7ba5b" gracePeriod=30 Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.298303 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-metadata" containerID="cri-o://4883f184578c131b7aea2b549436953ef6a4a1eb55f1d9a1053ca88e7faf8bfd" gracePeriod=30 Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.307683 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.139070612 podStartE2EDuration="7.307666201s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="2025-11-24 08:34:32.50638028 +0000 UTC m=+1146.381525433" lastFinishedPulling="2025-11-24 08:34:37.674975869 +0000 UTC m=+1151.550121022" observedRunningTime="2025-11-24 08:34:38.303890883 +0000 UTC m=+1152.179036056" watchObservedRunningTime="2025-11-24 08:34:38.307666201 +0000 UTC m=+1152.182811354" Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.309231 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" podStartSLOduration=5.309224795 podStartE2EDuration="5.309224795s" podCreationTimestamp="2025-11-24 08:34:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:37.284680949 +0000 UTC m=+1151.159826112" watchObservedRunningTime="2025-11-24 08:34:38.309224795 +0000 UTC m=+1152.184369948" Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.332782 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.6342160420000003 podStartE2EDuration="7.332760486s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="2025-11-24 08:34:32.986935194 +0000 UTC m=+1146.862080337" lastFinishedPulling="2025-11-24 08:34:37.685479628 +0000 UTC m=+1151.560624781" observedRunningTime="2025-11-24 08:34:38.327801325 +0000 UTC m=+1152.202946478" watchObservedRunningTime="2025-11-24 08:34:38.332760486 +0000 UTC m=+1152.207905639" Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.361944 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.671017711 podStartE2EDuration="7.361926528s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="2025-11-24 08:34:32.987264103 +0000 UTC m=+1146.862409256" lastFinishedPulling="2025-11-24 08:34:37.67817292 +0000 UTC m=+1151.553318073" observedRunningTime="2025-11-24 08:34:38.36130403 +0000 UTC m=+1152.236449193" watchObservedRunningTime="2025-11-24 08:34:38.361926528 +0000 UTC m=+1152.237071681" Nov 24 08:34:38 crc kubenswrapper[4831]: I1124 08:34:38.394616 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.757290591 podStartE2EDuration="7.394590559s" podCreationTimestamp="2025-11-24 08:34:31 +0000 UTC" firstStartedPulling="2025-11-24 08:34:33.04116547 +0000 UTC m=+1146.916310623" lastFinishedPulling="2025-11-24 08:34:37.678465448 +0000 UTC m=+1151.553610591" observedRunningTime="2025-11-24 08:34:38.387654552 +0000 UTC m=+1152.262799705" watchObservedRunningTime="2025-11-24 08:34:38.394590559 +0000 UTC m=+1152.269735732" Nov 24 08:34:39 crc kubenswrapper[4831]: I1124 08:34:39.310538 4831 generic.go:334] "Generic (PLEG): container finished" podID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerID="6ae768aa3b739a7e4daa33dc68d529f0a22bd59d873d4021bd75921fffd7ba5b" exitCode=143 Nov 24 08:34:39 crc kubenswrapper[4831]: I1124 08:34:39.310639 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerDied","Data":"6ae768aa3b739a7e4daa33dc68d529f0a22bd59d873d4021bd75921fffd7ba5b"} Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.779046 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.779561 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.815072 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.947810 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.947864 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:34:41 crc kubenswrapper[4831]: I1124 08:34:41.947880 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.212912 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.212955 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.365777 4831 generic.go:334] "Generic (PLEG): container finished" podID="8fbc0462-5e34-49ff-8786-5806fbbf36ec" containerID="e578b78210830b2e0a840fed543cf9da8fcdbf4715c4a33306b37635b129c57f" exitCode=0 Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.365848 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhjsw" event={"ID":"8fbc0462-5e34-49ff-8786-5806fbbf36ec","Type":"ContainerDied","Data":"e578b78210830b2e0a840fed543cf9da8fcdbf4715c4a33306b37635b129c57f"} Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.401003 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.610509 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.662869 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:34:42 crc kubenswrapper[4831]: I1124 08:34:42.663119 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="dnsmasq-dns" containerID="cri-o://56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0" gracePeriod=10 Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.032449 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.032626 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.171:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.186215 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.306898 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbsdt\" (UniqueName: \"kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt\") pod \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.307032 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb\") pod \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.307066 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb\") pod \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.307091 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config\") pod \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.307124 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc\") pod \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\" (UID: \"cfe5b851-71ed-43f4-b92a-2a3a913686b8\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.328895 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt" (OuterVolumeSpecName: "kube-api-access-rbsdt") pod "cfe5b851-71ed-43f4-b92a-2a3a913686b8" (UID: "cfe5b851-71ed-43f4-b92a-2a3a913686b8"). InnerVolumeSpecName "kube-api-access-rbsdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.372024 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cfe5b851-71ed-43f4-b92a-2a3a913686b8" (UID: "cfe5b851-71ed-43f4-b92a-2a3a913686b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.376268 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cfe5b851-71ed-43f4-b92a-2a3a913686b8" (UID: "cfe5b851-71ed-43f4-b92a-2a3a913686b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.384641 4831 generic.go:334] "Generic (PLEG): container finished" podID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerID="56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0" exitCode=0 Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.384900 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.385781 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" event={"ID":"cfe5b851-71ed-43f4-b92a-2a3a913686b8","Type":"ContainerDied","Data":"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0"} Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.385813 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-k9rn6" event={"ID":"cfe5b851-71ed-43f4-b92a-2a3a913686b8","Type":"ContainerDied","Data":"887875c472deb79b95b02e36292f298a2a1bde9f8b5e865f35e48898580df060"} Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.385847 4831 scope.go:117] "RemoveContainer" containerID="56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.409741 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config" (OuterVolumeSpecName: "config") pod "cfe5b851-71ed-43f4-b92a-2a3a913686b8" (UID: "cfe5b851-71ed-43f4-b92a-2a3a913686b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.411169 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbsdt\" (UniqueName: \"kubernetes.io/projected/cfe5b851-71ed-43f4-b92a-2a3a913686b8-kube-api-access-rbsdt\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.411230 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.411241 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.411273 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.412671 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cfe5b851-71ed-43f4-b92a-2a3a913686b8" (UID: "cfe5b851-71ed-43f4-b92a-2a3a913686b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.446798 4831 scope.go:117] "RemoveContainer" containerID="d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.480055 4831 scope.go:117] "RemoveContainer" containerID="56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0" Nov 24 08:34:43 crc kubenswrapper[4831]: E1124 08:34:43.480589 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0\": container with ID starting with 56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0 not found: ID does not exist" containerID="56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.480617 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0"} err="failed to get container status \"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0\": rpc error: code = NotFound desc = could not find container \"56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0\": container with ID starting with 56b2daca95a37be50c59f23958c88407301b578ba4188742028a40a160c2d0b0 not found: ID does not exist" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.480637 4831 scope.go:117] "RemoveContainer" containerID="d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855" Nov 24 08:34:43 crc kubenswrapper[4831]: E1124 08:34:43.481004 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855\": container with ID starting with d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855 not found: ID does not exist" containerID="d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.481033 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855"} err="failed to get container status \"d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855\": rpc error: code = NotFound desc = could not find container \"d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855\": container with ID starting with d7cb9a12d07c3294023d861f624fd429552969804636fad6b4515ff94944f855 not found: ID does not exist" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.518931 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfe5b851-71ed-43f4-b92a-2a3a913686b8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.687926 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.759975 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.767692 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-k9rn6"] Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.854200 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts\") pod \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.854257 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle\") pod \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.854343 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pzbz\" (UniqueName: \"kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz\") pod \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.854413 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data\") pod \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\" (UID: \"8fbc0462-5e34-49ff-8786-5806fbbf36ec\") " Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.870898 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts" (OuterVolumeSpecName: "scripts") pod "8fbc0462-5e34-49ff-8786-5806fbbf36ec" (UID: "8fbc0462-5e34-49ff-8786-5806fbbf36ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.874491 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz" (OuterVolumeSpecName: "kube-api-access-8pzbz") pod "8fbc0462-5e34-49ff-8786-5806fbbf36ec" (UID: "8fbc0462-5e34-49ff-8786-5806fbbf36ec"). InnerVolumeSpecName "kube-api-access-8pzbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.889055 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fbc0462-5e34-49ff-8786-5806fbbf36ec" (UID: "8fbc0462-5e34-49ff-8786-5806fbbf36ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.896423 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data" (OuterVolumeSpecName: "config-data") pod "8fbc0462-5e34-49ff-8786-5806fbbf36ec" (UID: "8fbc0462-5e34-49ff-8786-5806fbbf36ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.956715 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.956753 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.956763 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pzbz\" (UniqueName: \"kubernetes.io/projected/8fbc0462-5e34-49ff-8786-5806fbbf36ec-kube-api-access-8pzbz\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:43 crc kubenswrapper[4831]: I1124 08:34:43.956771 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbc0462-5e34-49ff-8786-5806fbbf36ec-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.398117 4831 generic.go:334] "Generic (PLEG): container finished" podID="2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" containerID="847c5e0f764c84d0674b9661873594c47402a7b00fbe858162979eae881f5ac9" exitCode=0 Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.398223 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" event={"ID":"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd","Type":"ContainerDied","Data":"847c5e0f764c84d0674b9661873594c47402a7b00fbe858162979eae881f5ac9"} Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.408545 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lhjsw" event={"ID":"8fbc0462-5e34-49ff-8786-5806fbbf36ec","Type":"ContainerDied","Data":"928a9696942d6a2d6dba4c02b1e10fef0fbff64ae29c4bef1d9ee8489547c45c"} Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.408603 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="928a9696942d6a2d6dba4c02b1e10fef0fbff64ae29c4bef1d9ee8489547c45c" Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.408688 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lhjsw" Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.575481 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.575697 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" containerName="nova-scheduler-scheduler" containerID="cri-o://8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2" gracePeriod=30 Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.622242 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.622580 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-log" containerID="cri-o://d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891" gracePeriod=30 Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.622892 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-api" containerID="cri-o://d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff" gracePeriod=30 Nov 24 08:34:44 crc kubenswrapper[4831]: I1124 08:34:44.903824 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" path="/var/lib/kubelet/pods/cfe5b851-71ed-43f4-b92a-2a3a913686b8/volumes" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.417266 4831 generic.go:334] "Generic (PLEG): container finished" podID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerID="d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891" exitCode=143 Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.417835 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerDied","Data":"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891"} Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.752932 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.827306 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data\") pod \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.827397 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle\") pod \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.827529 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts\") pod \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.827639 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7c89\" (UniqueName: \"kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89\") pod \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\" (UID: \"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd\") " Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.833409 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89" (OuterVolumeSpecName: "kube-api-access-h7c89") pod "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" (UID: "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd"). InnerVolumeSpecName "kube-api-access-h7c89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.879740 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts" (OuterVolumeSpecName: "scripts") pod "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" (UID: "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.931399 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.931600 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7c89\" (UniqueName: \"kubernetes.io/projected/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-kube-api-access-h7c89\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.933379 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data" (OuterVolumeSpecName: "config-data") pod "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" (UID: "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:45 crc kubenswrapper[4831]: I1124 08:34:45.951717 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" (UID: "2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.033193 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.033221 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.047388 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.133801 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data\") pod \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.133961 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgfw5\" (UniqueName: \"kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5\") pod \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.134055 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle\") pod \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\" (UID: \"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1\") " Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.139472 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5" (OuterVolumeSpecName: "kube-api-access-xgfw5") pod "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" (UID: "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1"). InnerVolumeSpecName "kube-api-access-xgfw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.164418 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" (UID: "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.164445 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data" (OuterVolumeSpecName: "config-data") pod "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" (UID: "d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.235183 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.235213 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgfw5\" (UniqueName: \"kubernetes.io/projected/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-kube-api-access-xgfw5\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.235223 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.431788 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.432758 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vl2xc" event={"ID":"2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd","Type":"ContainerDied","Data":"3445fbc41cd91e936065f75aac0f1c1d9680a4191369f85b9b9112ad3670f3b6"} Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.433027 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3445fbc41cd91e936065f75aac0f1c1d9680a4191369f85b9b9112ad3670f3b6" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.437945 4831 generic.go:334] "Generic (PLEG): container finished" podID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" containerID="8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2" exitCode=0 Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.438027 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1","Type":"ContainerDied","Data":"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2"} Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.438072 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1","Type":"ContainerDied","Data":"5964d8ee80bbeaccedac973ec82d23117b8bdd387308e64003000be2e75ddb56"} Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.438104 4831 scope.go:117] "RemoveContainer" containerID="8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.438366 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.476743 4831 scope.go:117] "RemoveContainer" containerID="8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2" Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.477329 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2\": container with ID starting with 8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2 not found: ID does not exist" containerID="8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.477443 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2"} err="failed to get container status \"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2\": rpc error: code = NotFound desc = could not find container \"8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2\": container with ID starting with 8a3ac1a7646dcf692fa823c36143aff3b759c62eb6069c885d3434eeec72efb2 not found: ID does not exist" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.480989 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.491397 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522393 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.522858 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" containerName="nova-cell1-conductor-db-sync" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522881 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" containerName="nova-cell1-conductor-db-sync" Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.522911 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="dnsmasq-dns" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522920 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="dnsmasq-dns" Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.522935 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbc0462-5e34-49ff-8786-5806fbbf36ec" containerName="nova-manage" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522943 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbc0462-5e34-49ff-8786-5806fbbf36ec" containerName="nova-manage" Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.522958 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="init" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522965 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="init" Nov 24 08:34:46 crc kubenswrapper[4831]: E1124 08:34:46.522976 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" containerName="nova-scheduler-scheduler" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.522983 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" containerName="nova-scheduler-scheduler" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.523198 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfe5b851-71ed-43f4-b92a-2a3a913686b8" containerName="dnsmasq-dns" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.523218 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbc0462-5e34-49ff-8786-5806fbbf36ec" containerName="nova-manage" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.523231 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" containerName="nova-scheduler-scheduler" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.523246 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" containerName="nova-cell1-conductor-db-sync" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.524029 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.530709 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.545200 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.545336 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.545384 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvrld\" (UniqueName: \"kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.559871 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.571262 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.572674 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.574514 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.600230 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.647311 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm8k7\" (UniqueName: \"kubernetes.io/projected/4aba5ba5-56d4-4883-af9e-9278046b3285-kube-api-access-xm8k7\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.647424 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.647456 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.648072 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvrld\" (UniqueName: \"kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.648137 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.648180 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.651263 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.651847 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.663183 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvrld\" (UniqueName: \"kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld\") pod \"nova-scheduler-0\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.750174 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.750516 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm8k7\" (UniqueName: \"kubernetes.io/projected/4aba5ba5-56d4-4883-af9e-9278046b3285-kube-api-access-xm8k7\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.750613 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.753436 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.754484 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4aba5ba5-56d4-4883-af9e-9278046b3285-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.765785 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm8k7\" (UniqueName: \"kubernetes.io/projected/4aba5ba5-56d4-4883-af9e-9278046b3285-kube-api-access-xm8k7\") pod \"nova-cell1-conductor-0\" (UID: \"4aba5ba5-56d4-4883-af9e-9278046b3285\") " pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.843416 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.891959 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:46 crc kubenswrapper[4831]: I1124 08:34:46.905912 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1" path="/var/lib/kubelet/pods/d246dd0b-f6b4-4fd0-a4b3-19e830cb38a1/volumes" Nov 24 08:34:47 crc kubenswrapper[4831]: I1124 08:34:47.360056 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:34:47 crc kubenswrapper[4831]: W1124 08:34:47.362626 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb518f57f_cc58_4b1e_be5e_74bc46526a4f.slice/crio-5d1973761458280e728b9e6dc145e3593d91e78f91ec4c7e3b98027711a490a8 WatchSource:0}: Error finding container 5d1973761458280e728b9e6dc145e3593d91e78f91ec4c7e3b98027711a490a8: Status 404 returned error can't find the container with id 5d1973761458280e728b9e6dc145e3593d91e78f91ec4c7e3b98027711a490a8 Nov 24 08:34:47 crc kubenswrapper[4831]: I1124 08:34:47.398438 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 24 08:34:47 crc kubenswrapper[4831]: W1124 08:34:47.417744 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4aba5ba5_56d4_4883_af9e_9278046b3285.slice/crio-7857c9651bcb44cca64156dc9d64ad5cfe0d2d8f7c4bd5698ba9dd0575bb0cbd WatchSource:0}: Error finding container 7857c9651bcb44cca64156dc9d64ad5cfe0d2d8f7c4bd5698ba9dd0575bb0cbd: Status 404 returned error can't find the container with id 7857c9651bcb44cca64156dc9d64ad5cfe0d2d8f7c4bd5698ba9dd0575bb0cbd Nov 24 08:34:47 crc kubenswrapper[4831]: I1124 08:34:47.450941 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4aba5ba5-56d4-4883-af9e-9278046b3285","Type":"ContainerStarted","Data":"7857c9651bcb44cca64156dc9d64ad5cfe0d2d8f7c4bd5698ba9dd0575bb0cbd"} Nov 24 08:34:47 crc kubenswrapper[4831]: I1124 08:34:47.456703 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b518f57f-cc58-4b1e-be5e-74bc46526a4f","Type":"ContainerStarted","Data":"5d1973761458280e728b9e6dc145e3593d91e78f91ec4c7e3b98027711a490a8"} Nov 24 08:34:48 crc kubenswrapper[4831]: I1124 08:34:48.470493 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b518f57f-cc58-4b1e-be5e-74bc46526a4f","Type":"ContainerStarted","Data":"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82"} Nov 24 08:34:48 crc kubenswrapper[4831]: I1124 08:34:48.475472 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4aba5ba5-56d4-4883-af9e-9278046b3285","Type":"ContainerStarted","Data":"2710f761322d1064658c1169c09cc242b09e677678ae236531c05225f9040637"} Nov 24 08:34:48 crc kubenswrapper[4831]: I1124 08:34:48.475705 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:48 crc kubenswrapper[4831]: I1124 08:34:48.503703 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.50367598 podStartE2EDuration="2.50367598s" podCreationTimestamp="2025-11-24 08:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:48.491560755 +0000 UTC m=+1162.366705908" watchObservedRunningTime="2025-11-24 08:34:48.50367598 +0000 UTC m=+1162.378821133" Nov 24 08:34:48 crc kubenswrapper[4831]: I1124 08:34:48.521435 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.521411026 podStartE2EDuration="2.521411026s" podCreationTimestamp="2025-11-24 08:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:48.519185403 +0000 UTC m=+1162.394330566" watchObservedRunningTime="2025-11-24 08:34:48.521411026 +0000 UTC m=+1162.396556199" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.490543 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.490639 4831 generic.go:334] "Generic (PLEG): container finished" podID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerID="d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff" exitCode=0 Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.490664 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerDied","Data":"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff"} Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.491006 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7a498f7-43b8-42c8-b2de-d006e6c2d787","Type":"ContainerDied","Data":"0bb563fe29b8dcd54175e1f192cfe8432f47dabe14b86ce91f78f88b4af5e364"} Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.491027 4831 scope.go:117] "RemoveContainer" containerID="d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.538744 4831 scope.go:117] "RemoveContainer" containerID="d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.568407 4831 scope.go:117] "RemoveContainer" containerID="d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff" Nov 24 08:34:49 crc kubenswrapper[4831]: E1124 08:34:49.571016 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff\": container with ID starting with d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff not found: ID does not exist" containerID="d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.571084 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff"} err="failed to get container status \"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff\": rpc error: code = NotFound desc = could not find container \"d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff\": container with ID starting with d9c21f1cfd135abc0d6ed6e7c6db7292b302996ca878cfe7050b7503198319ff not found: ID does not exist" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.571143 4831 scope.go:117] "RemoveContainer" containerID="d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891" Nov 24 08:34:49 crc kubenswrapper[4831]: E1124 08:34:49.571609 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891\": container with ID starting with d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891 not found: ID does not exist" containerID="d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.571653 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891"} err="failed to get container status \"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891\": rpc error: code = NotFound desc = could not find container \"d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891\": container with ID starting with d4054653a27bb59501e521c058ed069ceee7044e38c90e9a8d0816ffa2a7a891 not found: ID does not exist" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.611801 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle\") pod \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.611917 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs\") pod \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.612005 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data\") pod \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.612035 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjh54\" (UniqueName: \"kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54\") pod \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\" (UID: \"a7a498f7-43b8-42c8-b2de-d006e6c2d787\") " Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.613241 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs" (OuterVolumeSpecName: "logs") pod "a7a498f7-43b8-42c8-b2de-d006e6c2d787" (UID: "a7a498f7-43b8-42c8-b2de-d006e6c2d787"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.617081 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54" (OuterVolumeSpecName: "kube-api-access-wjh54") pod "a7a498f7-43b8-42c8-b2de-d006e6c2d787" (UID: "a7a498f7-43b8-42c8-b2de-d006e6c2d787"). InnerVolumeSpecName "kube-api-access-wjh54". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.640029 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data" (OuterVolumeSpecName: "config-data") pod "a7a498f7-43b8-42c8-b2de-d006e6c2d787" (UID: "a7a498f7-43b8-42c8-b2de-d006e6c2d787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.668404 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7a498f7-43b8-42c8-b2de-d006e6c2d787" (UID: "a7a498f7-43b8-42c8-b2de-d006e6c2d787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.713568 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.713601 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7a498f7-43b8-42c8-b2de-d006e6c2d787-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.713611 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a498f7-43b8-42c8-b2de-d006e6c2d787-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:49 crc kubenswrapper[4831]: I1124 08:34:49.713619 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjh54\" (UniqueName: \"kubernetes.io/projected/a7a498f7-43b8-42c8-b2de-d006e6c2d787-kube-api-access-wjh54\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.500414 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.553508 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.579462 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.599940 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:50 crc kubenswrapper[4831]: E1124 08:34:50.600484 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-api" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.600503 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-api" Nov 24 08:34:50 crc kubenswrapper[4831]: E1124 08:34:50.600540 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-log" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.600549 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-log" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.600789 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-api" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.600814 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" containerName="nova-api-log" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.602094 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.604650 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.626106 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.632521 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.632581 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.632642 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glmch\" (UniqueName: \"kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.632787 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.734247 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.734303 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.734388 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.734422 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glmch\" (UniqueName: \"kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.734964 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.738596 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.748925 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.757254 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glmch\" (UniqueName: \"kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch\") pod \"nova-api-0\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " pod="openstack/nova-api-0" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.903246 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a498f7-43b8-42c8-b2de-d006e6c2d787" path="/var/lib/kubelet/pods/a7a498f7-43b8-42c8-b2de-d006e6c2d787/volumes" Nov 24 08:34:50 crc kubenswrapper[4831]: I1124 08:34:50.950722 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:34:51 crc kubenswrapper[4831]: I1124 08:34:51.460674 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:34:51 crc kubenswrapper[4831]: W1124 08:34:51.467569 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod503f7d3b_71dc_4fc1_b2be_858dd654db32.slice/crio-82f19cb8e4c0454dd3996cb52f8c09cabd237a9f89e50bb9f217e4ab2e164e71 WatchSource:0}: Error finding container 82f19cb8e4c0454dd3996cb52f8c09cabd237a9f89e50bb9f217e4ab2e164e71: Status 404 returned error can't find the container with id 82f19cb8e4c0454dd3996cb52f8c09cabd237a9f89e50bb9f217e4ab2e164e71 Nov 24 08:34:51 crc kubenswrapper[4831]: I1124 08:34:51.513434 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerStarted","Data":"82f19cb8e4c0454dd3996cb52f8c09cabd237a9f89e50bb9f217e4ab2e164e71"} Nov 24 08:34:51 crc kubenswrapper[4831]: I1124 08:34:51.844393 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:34:52 crc kubenswrapper[4831]: I1124 08:34:52.259232 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 08:34:52 crc kubenswrapper[4831]: I1124 08:34:52.532717 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerStarted","Data":"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9"} Nov 24 08:34:52 crc kubenswrapper[4831]: I1124 08:34:52.532761 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerStarted","Data":"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be"} Nov 24 08:34:52 crc kubenswrapper[4831]: I1124 08:34:52.551566 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.55154145 podStartE2EDuration="2.55154145s" podCreationTimestamp="2025-11-24 08:34:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:34:52.548140683 +0000 UTC m=+1166.423285866" watchObservedRunningTime="2025-11-24 08:34:52.55154145 +0000 UTC m=+1166.426686643" Nov 24 08:34:54 crc kubenswrapper[4831]: I1124 08:34:54.810018 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:54 crc kubenswrapper[4831]: I1124 08:34:54.810275 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3294a544-5049-47a8-8633-2250f41005ea" containerName="kube-state-metrics" containerID="cri-o://cd2098e6b3f498379b7b998f5714d4e29a6764be559af2a2f95b22ad2487511a" gracePeriod=30 Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.559099 4831 generic.go:334] "Generic (PLEG): container finished" podID="3294a544-5049-47a8-8633-2250f41005ea" containerID="cd2098e6b3f498379b7b998f5714d4e29a6764be559af2a2f95b22ad2487511a" exitCode=2 Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.559177 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3294a544-5049-47a8-8633-2250f41005ea","Type":"ContainerDied","Data":"cd2098e6b3f498379b7b998f5714d4e29a6764be559af2a2f95b22ad2487511a"} Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.559793 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3294a544-5049-47a8-8633-2250f41005ea","Type":"ContainerDied","Data":"9f93f8fb4442131a65010e9afa13bc22e26786a6ab33643220df40f176757af2"} Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.559813 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f93f8fb4442131a65010e9afa13bc22e26786a6ab33643220df40f176757af2" Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.616647 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.682866 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkhlm\" (UniqueName: \"kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm\") pod \"3294a544-5049-47a8-8633-2250f41005ea\" (UID: \"3294a544-5049-47a8-8633-2250f41005ea\") " Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.696613 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm" (OuterVolumeSpecName: "kube-api-access-wkhlm") pod "3294a544-5049-47a8-8633-2250f41005ea" (UID: "3294a544-5049-47a8-8633-2250f41005ea"). InnerVolumeSpecName "kube-api-access-wkhlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:34:55 crc kubenswrapper[4831]: I1124 08:34:55.785555 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkhlm\" (UniqueName: \"kubernetes.io/projected/3294a544-5049-47a8-8633-2250f41005ea-kube-api-access-wkhlm\") on node \"crc\" DevicePath \"\"" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.566693 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.596689 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.612632 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.620299 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:56 crc kubenswrapper[4831]: E1124 08:34:56.620804 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3294a544-5049-47a8-8633-2250f41005ea" containerName="kube-state-metrics" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.620827 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="3294a544-5049-47a8-8633-2250f41005ea" containerName="kube-state-metrics" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.621062 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="3294a544-5049-47a8-8633-2250f41005ea" containerName="kube-state-metrics" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.621796 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.626106 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.626586 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.632445 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.708754 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.708947 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.709110 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.709140 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47tn\" (UniqueName: \"kubernetes.io/projected/76ead5ac-ef11-4309-b031-10ea621a212e-kube-api-access-s47tn\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.811198 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.811250 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47tn\" (UniqueName: \"kubernetes.io/projected/76ead5ac-ef11-4309-b031-10ea621a212e-kube-api-access-s47tn\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.811327 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.811444 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.817282 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.821295 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.837737 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.838309 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-central-agent" containerID="cri-o://a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793" gracePeriod=30 Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.838474 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47tn\" (UniqueName: \"kubernetes.io/projected/76ead5ac-ef11-4309-b031-10ea621a212e-kube-api-access-s47tn\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.838614 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="proxy-httpd" containerID="cri-o://32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d" gracePeriod=30 Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.838760 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="sg-core" containerID="cri-o://61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb" gracePeriod=30 Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.838884 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-notification-agent" containerID="cri-o://60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35" gracePeriod=30 Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.841226 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/76ead5ac-ef11-4309-b031-10ea621a212e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"76ead5ac-ef11-4309-b031-10ea621a212e\") " pod="openstack/kube-state-metrics-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.844637 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.907689 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3294a544-5049-47a8-8633-2250f41005ea" path="/var/lib/kubelet/pods/3294a544-5049-47a8-8633-2250f41005ea/volumes" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.923192 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.932979 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 24 08:34:56 crc kubenswrapper[4831]: I1124 08:34:56.947388 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.509676 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.531856 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576172 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerDied","Data":"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d"} Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576287 4831 generic.go:334] "Generic (PLEG): container finished" podID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerID="32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d" exitCode=0 Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576331 4831 generic.go:334] "Generic (PLEG): container finished" podID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerID="61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb" exitCode=2 Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576339 4831 generic.go:334] "Generic (PLEG): container finished" podID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerID="a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793" exitCode=0 Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576381 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerDied","Data":"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb"} Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.576406 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerDied","Data":"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793"} Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.584898 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76ead5ac-ef11-4309-b031-10ea621a212e","Type":"ContainerStarted","Data":"df43eeb3dbcb5485f1954044498da160c4498c92f7b4df1a59bec858be219a18"} Nov 24 08:34:57 crc kubenswrapper[4831]: I1124 08:34:57.620155 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:34:58 crc kubenswrapper[4831]: I1124 08:34:58.595503 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"76ead5ac-ef11-4309-b031-10ea621a212e","Type":"ContainerStarted","Data":"0484a4333e92f0ac2724ead870270143ed52347f9621de28a178364e36e94029"} Nov 24 08:34:58 crc kubenswrapper[4831]: I1124 08:34:58.596559 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 24 08:34:58 crc kubenswrapper[4831]: I1124 08:34:58.626278 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.180641809 podStartE2EDuration="2.626256746s" podCreationTimestamp="2025-11-24 08:34:56 +0000 UTC" firstStartedPulling="2025-11-24 08:34:57.531532999 +0000 UTC m=+1171.406678152" lastFinishedPulling="2025-11-24 08:34:57.977147936 +0000 UTC m=+1171.852293089" observedRunningTime="2025-11-24 08:34:58.623286992 +0000 UTC m=+1172.498432175" watchObservedRunningTime="2025-11-24 08:34:58.626256746 +0000 UTC m=+1172.501401909" Nov 24 08:35:00 crc kubenswrapper[4831]: I1124 08:35:00.951359 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:00 crc kubenswrapper[4831]: I1124 08:35:00.951721 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.034613 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.035256 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.559474 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616191 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616271 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616352 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616387 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616450 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616525 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p55bm\" (UniqueName: \"kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.616574 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd\") pod \"5f0d83ce-790d-48d2-922b-0db305a983a1\" (UID: \"5f0d83ce-790d-48d2-922b-0db305a983a1\") " Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.617191 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.617952 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.623445 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm" (OuterVolumeSpecName: "kube-api-access-p55bm") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "kube-api-access-p55bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.651514 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts" (OuterVolumeSpecName: "scripts") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.654229 4831 generic.go:334] "Generic (PLEG): container finished" podID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerID="60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35" exitCode=0 Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.654267 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerDied","Data":"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35"} Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.654298 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5f0d83ce-790d-48d2-922b-0db305a983a1","Type":"ContainerDied","Data":"e53c1c5f14f7137a08016d2cfa0db28f8bb70b00d2e422dc863198f1272c1293"} Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.654336 4831 scope.go:117] "RemoveContainer" containerID="32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.654480 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.662061 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.715855 4831 scope.go:117] "RemoveContainer" containerID="61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.718305 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.718558 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p55bm\" (UniqueName: \"kubernetes.io/projected/5f0d83ce-790d-48d2-922b-0db305a983a1-kube-api-access-p55bm\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.718606 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.718617 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.718626 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5f0d83ce-790d-48d2-922b-0db305a983a1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.720532 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.738988 4831 scope.go:117] "RemoveContainer" containerID="60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.760530 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data" (OuterVolumeSpecName: "config-data") pod "5f0d83ce-790d-48d2-922b-0db305a983a1" (UID: "5f0d83ce-790d-48d2-922b-0db305a983a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.782981 4831 scope.go:117] "RemoveContainer" containerID="a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.800181 4831 scope.go:117] "RemoveContainer" containerID="32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d" Nov 24 08:35:02 crc kubenswrapper[4831]: E1124 08:35:02.801528 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d\": container with ID starting with 32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d not found: ID does not exist" containerID="32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.801567 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d"} err="failed to get container status \"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d\": rpc error: code = NotFound desc = could not find container \"32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d\": container with ID starting with 32fe3ba1b785e94f8b971245e65b2bac8246738c2da5e90570d60b931062f01d not found: ID does not exist" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.801590 4831 scope.go:117] "RemoveContainer" containerID="61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb" Nov 24 08:35:02 crc kubenswrapper[4831]: E1124 08:35:02.801979 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb\": container with ID starting with 61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb not found: ID does not exist" containerID="61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.802005 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb"} err="failed to get container status \"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb\": rpc error: code = NotFound desc = could not find container \"61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb\": container with ID starting with 61e083137715f90944b51298958911ea0e923a2e38ebe5d536276a5aaf8df2bb not found: ID does not exist" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.802022 4831 scope.go:117] "RemoveContainer" containerID="60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35" Nov 24 08:35:02 crc kubenswrapper[4831]: E1124 08:35:02.802300 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35\": container with ID starting with 60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35 not found: ID does not exist" containerID="60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.802354 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35"} err="failed to get container status \"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35\": rpc error: code = NotFound desc = could not find container \"60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35\": container with ID starting with 60169daa30c3a264fd2053419b90a9ab6e3e59dd006d00d32bbf50638c8e5f35 not found: ID does not exist" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.802369 4831 scope.go:117] "RemoveContainer" containerID="a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793" Nov 24 08:35:02 crc kubenswrapper[4831]: E1124 08:35:02.802645 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793\": container with ID starting with a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793 not found: ID does not exist" containerID="a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.802668 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793"} err="failed to get container status \"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793\": rpc error: code = NotFound desc = could not find container \"a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793\": container with ID starting with a87f54bc7916fb51d7c9d3fb64a93f389263dd09f410cc06fbac9ba660e28793 not found: ID does not exist" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.820525 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.820556 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f0d83ce-790d-48d2-922b-0db305a983a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.982213 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:02 crc kubenswrapper[4831]: I1124 08:35:02.992478 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013351 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:03 crc kubenswrapper[4831]: E1124 08:35:03.013681 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="proxy-httpd" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013696 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="proxy-httpd" Nov 24 08:35:03 crc kubenswrapper[4831]: E1124 08:35:03.013725 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-central-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013731 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-central-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: E1124 08:35:03.013741 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="sg-core" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013746 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="sg-core" Nov 24 08:35:03 crc kubenswrapper[4831]: E1124 08:35:03.013758 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-notification-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013764 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-notification-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013922 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="sg-core" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013932 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="proxy-httpd" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013945 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-central-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.013951 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" containerName="ceilometer-notification-agent" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.016186 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.018569 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.018874 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.028671 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.038161 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124681 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124770 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124806 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vwz\" (UniqueName: \"kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124876 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124917 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124937 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124984 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.124999 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.226690 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227294 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227360 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227426 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227450 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vwz\" (UniqueName: \"kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227514 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227551 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227572 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.227967 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.228786 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.231353 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.231878 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.233221 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.233811 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.241632 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.255051 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vwz\" (UniqueName: \"kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz\") pod \"ceilometer-0\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.333572 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:03 crc kubenswrapper[4831]: I1124 08:35:03.919179 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:04 crc kubenswrapper[4831]: I1124 08:35:04.675903 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerStarted","Data":"06d7c68d63ab81a63c920c91d84c612208731f81ac906c1389468f8f37e58543"} Nov 24 08:35:04 crc kubenswrapper[4831]: I1124 08:35:04.906456 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f0d83ce-790d-48d2-922b-0db305a983a1" path="/var/lib/kubelet/pods/5f0d83ce-790d-48d2-922b-0db305a983a1/volumes" Nov 24 08:35:05 crc kubenswrapper[4831]: I1124 08:35:05.686190 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerStarted","Data":"731b35038551ad9a20c727df836abceeafb2bd717d22e50e900e09f2fa17c76c"} Nov 24 08:35:05 crc kubenswrapper[4831]: I1124 08:35:05.686462 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerStarted","Data":"c82cbca90d6dbf2001e09629543bb99cf200256cf60b37c4049dd6a5d84490c4"} Nov 24 08:35:06 crc kubenswrapper[4831]: I1124 08:35:06.695025 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerStarted","Data":"bbc6e9020ce2607361b81c7b0534c737271013b3a9c1ab457effbcb331441afe"} Nov 24 08:35:06 crc kubenswrapper[4831]: I1124 08:35:06.958915 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 24 08:35:07 crc kubenswrapper[4831]: I1124 08:35:07.704227 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerStarted","Data":"3c8adc5c24b1e6007e3145f67904f93653b69c7e8dc9e5714d96e15ab3ef6d20"} Nov 24 08:35:07 crc kubenswrapper[4831]: I1124 08:35:07.704606 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.713449 4831 generic.go:334] "Generic (PLEG): container finished" podID="9d3bf305-0549-4809-acc9-3ea794deec4c" containerID="c64e6fd12cf514241c834e3d9f449c2e4758b32856e47c5d2b4db7566f5c0cdd" exitCode=137 Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.713628 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9d3bf305-0549-4809-acc9-3ea794deec4c","Type":"ContainerDied","Data":"c64e6fd12cf514241c834e3d9f449c2e4758b32856e47c5d2b4db7566f5c0cdd"} Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.714173 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9d3bf305-0549-4809-acc9-3ea794deec4c","Type":"ContainerDied","Data":"a970a13b929df4590e2f3ef618fb1eb6049f932948d57aa3baaba3da1a7d17ae"} Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.714191 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a970a13b929df4590e2f3ef618fb1eb6049f932948d57aa3baaba3da1a7d17ae" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.715543 4831 generic.go:334] "Generic (PLEG): container finished" podID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerID="4883f184578c131b7aea2b549436953ef6a4a1eb55f1d9a1053ca88e7faf8bfd" exitCode=137 Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.715774 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerDied","Data":"4883f184578c131b7aea2b549436953ef6a4a1eb55f1d9a1053ca88e7faf8bfd"} Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.715795 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"217fa2be-0216-4d8a-929e-1c45e083c00b","Type":"ContainerDied","Data":"c13bac6e70f7fdcce1c2938098b3e56825d5511fccf3aca34bf9268b75b2a2b6"} Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.715804 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c13bac6e70f7fdcce1c2938098b3e56825d5511fccf3aca34bf9268b75b2a2b6" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.779827 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.782603 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.795354 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.402711455 podStartE2EDuration="6.795339189s" podCreationTimestamp="2025-11-24 08:35:02 +0000 UTC" firstStartedPulling="2025-11-24 08:35:03.917088441 +0000 UTC m=+1177.792233594" lastFinishedPulling="2025-11-24 08:35:07.309716175 +0000 UTC m=+1181.184861328" observedRunningTime="2025-11-24 08:35:07.73351269 +0000 UTC m=+1181.608657853" watchObservedRunningTime="2025-11-24 08:35:08.795339189 +0000 UTC m=+1182.670484342" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831435 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lhw5\" (UniqueName: \"kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5\") pod \"9d3bf305-0549-4809-acc9-3ea794deec4c\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831491 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data\") pod \"217fa2be-0216-4d8a-929e-1c45e083c00b\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831527 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle\") pod \"217fa2be-0216-4d8a-929e-1c45e083c00b\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831573 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data\") pod \"9d3bf305-0549-4809-acc9-3ea794deec4c\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831644 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle\") pod \"9d3bf305-0549-4809-acc9-3ea794deec4c\" (UID: \"9d3bf305-0549-4809-acc9-3ea794deec4c\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831713 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs\") pod \"217fa2be-0216-4d8a-929e-1c45e083c00b\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.831779 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c468q\" (UniqueName: \"kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q\") pod \"217fa2be-0216-4d8a-929e-1c45e083c00b\" (UID: \"217fa2be-0216-4d8a-929e-1c45e083c00b\") " Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.851046 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs" (OuterVolumeSpecName: "logs") pod "217fa2be-0216-4d8a-929e-1c45e083c00b" (UID: "217fa2be-0216-4d8a-929e-1c45e083c00b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.853470 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q" (OuterVolumeSpecName: "kube-api-access-c468q") pod "217fa2be-0216-4d8a-929e-1c45e083c00b" (UID: "217fa2be-0216-4d8a-929e-1c45e083c00b"). InnerVolumeSpecName "kube-api-access-c468q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.864157 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5" (OuterVolumeSpecName: "kube-api-access-8lhw5") pod "9d3bf305-0549-4809-acc9-3ea794deec4c" (UID: "9d3bf305-0549-4809-acc9-3ea794deec4c"). InnerVolumeSpecName "kube-api-access-8lhw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.865673 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "217fa2be-0216-4d8a-929e-1c45e083c00b" (UID: "217fa2be-0216-4d8a-929e-1c45e083c00b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.907906 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data" (OuterVolumeSpecName: "config-data") pod "9d3bf305-0549-4809-acc9-3ea794deec4c" (UID: "9d3bf305-0549-4809-acc9-3ea794deec4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.909845 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d3bf305-0549-4809-acc9-3ea794deec4c" (UID: "9d3bf305-0549-4809-acc9-3ea794deec4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.919401 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data" (OuterVolumeSpecName: "config-data") pod "217fa2be-0216-4d8a-929e-1c45e083c00b" (UID: "217fa2be-0216-4d8a-929e-1c45e083c00b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934191 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934405 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3bf305-0549-4809-acc9-3ea794deec4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934485 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/217fa2be-0216-4d8a-929e-1c45e083c00b-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934544 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c468q\" (UniqueName: \"kubernetes.io/projected/217fa2be-0216-4d8a-929e-1c45e083c00b-kube-api-access-c468q\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934605 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lhw5\" (UniqueName: \"kubernetes.io/projected/9d3bf305-0549-4809-acc9-3ea794deec4c-kube-api-access-8lhw5\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934659 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:08 crc kubenswrapper[4831]: I1124 08:35:08.934724 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/217fa2be-0216-4d8a-929e-1c45e083c00b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.726202 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.726214 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.758455 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.772017 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.789801 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.803967 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.813550 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: E1124 08:35:09.814021 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-log" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814044 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-log" Nov 24 08:35:09 crc kubenswrapper[4831]: E1124 08:35:09.814063 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-metadata" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814072 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-metadata" Nov 24 08:35:09 crc kubenswrapper[4831]: E1124 08:35:09.814111 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3bf305-0549-4809-acc9-3ea794deec4c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814121 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3bf305-0549-4809-acc9-3ea794deec4c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814381 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-metadata" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814413 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" containerName="nova-metadata-log" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.814430 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3bf305-0549-4809-acc9-3ea794deec4c" containerName="nova-cell1-novncproxy-novncproxy" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.815000 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.817762 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.819731 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.819831 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.832117 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.832201 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.833761 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.838284 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.838521 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.855675 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949212 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949255 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949377 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb6b\" (UniqueName: \"kubernetes.io/projected/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-kube-api-access-rwb6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949411 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949463 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949489 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgvfd\" (UniqueName: \"kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949545 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949563 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949586 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:09 crc kubenswrapper[4831]: I1124 08:35:09.949679 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051475 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051537 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051600 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051624 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051711 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb6b\" (UniqueName: \"kubernetes.io/projected/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-kube-api-access-rwb6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051754 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051792 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051818 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgvfd\" (UniqueName: \"kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051868 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.051894 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.052896 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.058167 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.058177 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.060834 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.061025 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.062476 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.069852 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb6b\" (UniqueName: \"kubernetes.io/projected/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-kube-api-access-rwb6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.073952 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.076355 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4112313c-5cfc-4e0a-bf38-cd6cf34f309e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4112313c-5cfc-4e0a-bf38-cd6cf34f309e\") " pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.076912 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgvfd\" (UniqueName: \"kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd\") pod \"nova-metadata-0\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.145524 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.167771 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.606837 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.753883 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4112313c-5cfc-4e0a-bf38-cd6cf34f309e","Type":"ContainerStarted","Data":"2e7309f7506a75bfdc48e2c54866009906866ba8beeab4065d404216f04e7573"} Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.798921 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:10 crc kubenswrapper[4831]: W1124 08:35:10.830062 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b8c6a67_2bd5_403d_8c0a_6adb179b0d04.slice/crio-da7f53de936bc12546b7e100116fffa99ec9e4e0daaf7ff219ae2384f22beeb1 WatchSource:0}: Error finding container da7f53de936bc12546b7e100116fffa99ec9e4e0daaf7ff219ae2384f22beeb1: Status 404 returned error can't find the container with id da7f53de936bc12546b7e100116fffa99ec9e4e0daaf7ff219ae2384f22beeb1 Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.903830 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="217fa2be-0216-4d8a-929e-1c45e083c00b" path="/var/lib/kubelet/pods/217fa2be-0216-4d8a-929e-1c45e083c00b/volumes" Nov 24 08:35:10 crc kubenswrapper[4831]: I1124 08:35:10.904528 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3bf305-0549-4809-acc9-3ea794deec4c" path="/var/lib/kubelet/pods/9d3bf305-0549-4809-acc9-3ea794deec4c/volumes" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.016138 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.016220 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.017146 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.017187 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.027274 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.030027 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.268998 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.278208 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.292105 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.404271 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.404352 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.404415 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.404474 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lll9p\" (UniqueName: \"kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.404499 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.506253 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.506305 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.506376 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.506440 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lll9p\" (UniqueName: \"kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.506465 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.507453 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.507516 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.508149 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.508693 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.525065 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lll9p\" (UniqueName: \"kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p\") pod \"dnsmasq-dns-68d4b6d797-lrjz9\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.602931 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.776129 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4112313c-5cfc-4e0a-bf38-cd6cf34f309e","Type":"ContainerStarted","Data":"422178695a836875166adca5031373f328baa4b5a69c2a24ce97c77d6d6e323b"} Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.804369 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerStarted","Data":"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a"} Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.804426 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerStarted","Data":"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7"} Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.804438 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerStarted","Data":"da7f53de936bc12546b7e100116fffa99ec9e4e0daaf7ff219ae2384f22beeb1"} Nov 24 08:35:11 crc kubenswrapper[4831]: I1124 08:35:11.846884 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.8468667659999998 podStartE2EDuration="2.846866766s" podCreationTimestamp="2025-11-24 08:35:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:11.819722772 +0000 UTC m=+1185.694867945" watchObservedRunningTime="2025-11-24 08:35:11.846866766 +0000 UTC m=+1185.722011919" Nov 24 08:35:12 crc kubenswrapper[4831]: I1124 08:35:12.244597 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.244581587 podStartE2EDuration="3.244581587s" podCreationTimestamp="2025-11-24 08:35:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:11.840619838 +0000 UTC m=+1185.715764991" watchObservedRunningTime="2025-11-24 08:35:12.244581587 +0000 UTC m=+1186.119726740" Nov 24 08:35:12 crc kubenswrapper[4831]: I1124 08:35:12.249713 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:35:12 crc kubenswrapper[4831]: W1124 08:35:12.260941 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21835b63_320c_4149_b01d_d7afa343a341.slice/crio-81d1729848168c7b9f50f6b136e85212dfa2e1f13014b08a9f9f33510c47450a WatchSource:0}: Error finding container 81d1729848168c7b9f50f6b136e85212dfa2e1f13014b08a9f9f33510c47450a: Status 404 returned error can't find the container with id 81d1729848168c7b9f50f6b136e85212dfa2e1f13014b08a9f9f33510c47450a Nov 24 08:35:12 crc kubenswrapper[4831]: I1124 08:35:12.813304 4831 generic.go:334] "Generic (PLEG): container finished" podID="21835b63-320c-4149-b01d-d7afa343a341" containerID="42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469" exitCode=0 Nov 24 08:35:12 crc kubenswrapper[4831]: I1124 08:35:12.813440 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" event={"ID":"21835b63-320c-4149-b01d-d7afa343a341","Type":"ContainerDied","Data":"42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469"} Nov 24 08:35:12 crc kubenswrapper[4831]: I1124 08:35:12.813465 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" event={"ID":"21835b63-320c-4149-b01d-d7afa343a341","Type":"ContainerStarted","Data":"81d1729848168c7b9f50f6b136e85212dfa2e1f13014b08a9f9f33510c47450a"} Nov 24 08:35:13 crc kubenswrapper[4831]: I1124 08:35:13.822829 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" event={"ID":"21835b63-320c-4149-b01d-d7afa343a341","Type":"ContainerStarted","Data":"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772"} Nov 24 08:35:13 crc kubenswrapper[4831]: I1124 08:35:13.823160 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.335673 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" podStartSLOduration=3.335655997 podStartE2EDuration="3.335655997s" podCreationTimestamp="2025-11-24 08:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:13.860214899 +0000 UTC m=+1187.735360062" watchObservedRunningTime="2025-11-24 08:35:14.335655997 +0000 UTC m=+1188.210801150" Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.342662 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.342928 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-central-agent" containerID="cri-o://c82cbca90d6dbf2001e09629543bb99cf200256cf60b37c4049dd6a5d84490c4" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.342983 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-notification-agent" containerID="cri-o://731b35038551ad9a20c727df836abceeafb2bd717d22e50e900e09f2fa17c76c" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.342990 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="sg-core" containerID="cri-o://bbc6e9020ce2607361b81c7b0534c737271013b3a9c1ab457effbcb331441afe" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.343174 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="proxy-httpd" containerID="cri-o://3c8adc5c24b1e6007e3145f67904f93653b69c7e8dc9e5714d96e15ab3ef6d20" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.825662 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.825917 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-log" containerID="cri-o://d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.826217 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-api" containerID="cri-o://9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9" gracePeriod=30 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.856413 4831 generic.go:334] "Generic (PLEG): container finished" podID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerID="3c8adc5c24b1e6007e3145f67904f93653b69c7e8dc9e5714d96e15ab3ef6d20" exitCode=0 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.856756 4831 generic.go:334] "Generic (PLEG): container finished" podID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerID="bbc6e9020ce2607361b81c7b0534c737271013b3a9c1ab457effbcb331441afe" exitCode=2 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.856767 4831 generic.go:334] "Generic (PLEG): container finished" podID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerID="731b35038551ad9a20c727df836abceeafb2bd717d22e50e900e09f2fa17c76c" exitCode=0 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.856778 4831 generic.go:334] "Generic (PLEG): container finished" podID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerID="c82cbca90d6dbf2001e09629543bb99cf200256cf60b37c4049dd6a5d84490c4" exitCode=0 Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.857711 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerDied","Data":"3c8adc5c24b1e6007e3145f67904f93653b69c7e8dc9e5714d96e15ab3ef6d20"} Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.857746 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerDied","Data":"bbc6e9020ce2607361b81c7b0534c737271013b3a9c1ab457effbcb331441afe"} Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.857762 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerDied","Data":"731b35038551ad9a20c727df836abceeafb2bd717d22e50e900e09f2fa17c76c"} Nov 24 08:35:14 crc kubenswrapper[4831]: I1124 08:35:14.857775 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerDied","Data":"c82cbca90d6dbf2001e09629543bb99cf200256cf60b37c4049dd6a5d84490c4"} Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.152479 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.170571 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.172271 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.226758 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378179 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378281 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378300 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378331 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378413 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378435 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378495 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.378545 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84vwz\" (UniqueName: \"kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz\") pod \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\" (UID: \"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6\") " Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.379379 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.379947 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.384155 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz" (OuterVolumeSpecName: "kube-api-access-84vwz") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "kube-api-access-84vwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.384253 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts" (OuterVolumeSpecName: "scripts") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.407552 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.435853 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.451795 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480310 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480357 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480373 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84vwz\" (UniqueName: \"kubernetes.io/projected/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-kube-api-access-84vwz\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480387 4831 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480397 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480408 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.480418 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.488339 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data" (OuterVolumeSpecName: "config-data") pod "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" (UID: "18c303ee-4bdf-4d13-9dc7-26c78fba7ce6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.582170 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.866270 4831 generic.go:334] "Generic (PLEG): container finished" podID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerID="d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be" exitCode=143 Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.866420 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerDied","Data":"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be"} Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.870525 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18c303ee-4bdf-4d13-9dc7-26c78fba7ce6","Type":"ContainerDied","Data":"06d7c68d63ab81a63c920c91d84c612208731f81ac906c1389468f8f37e58543"} Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.870575 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.870577 4831 scope.go:117] "RemoveContainer" containerID="3c8adc5c24b1e6007e3145f67904f93653b69c7e8dc9e5714d96e15ab3ef6d20" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.923370 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.934205 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.934578 4831 scope.go:117] "RemoveContainer" containerID="bbc6e9020ce2607361b81c7b0534c737271013b3a9c1ab457effbcb331441afe" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948302 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:15 crc kubenswrapper[4831]: E1124 08:35:15.948716 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-notification-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948733 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-notification-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: E1124 08:35:15.948747 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="proxy-httpd" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948754 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="proxy-httpd" Nov 24 08:35:15 crc kubenswrapper[4831]: E1124 08:35:15.948768 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-central-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948775 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-central-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: E1124 08:35:15.948808 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="sg-core" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948816 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="sg-core" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.948976 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="sg-core" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.949000 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="proxy-httpd" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.949016 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-notification-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.949024 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" containerName="ceilometer-central-agent" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.953707 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.959813 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.960019 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.960161 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.967600 4831 scope.go:117] "RemoveContainer" containerID="731b35038551ad9a20c727df836abceeafb2bd717d22e50e900e09f2fa17c76c" Nov 24 08:35:15 crc kubenswrapper[4831]: I1124 08:35:15.968811 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.006053 4831 scope.go:117] "RemoveContainer" containerID="c82cbca90d6dbf2001e09629543bb99cf200256cf60b37c4049dd6a5d84490c4" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092408 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092488 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092516 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092610 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4ncp\" (UniqueName: \"kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092651 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092689 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092786 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.092962 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194236 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194288 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194346 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4ncp\" (UniqueName: \"kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194366 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194397 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194417 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194461 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194522 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.194957 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.195059 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.200048 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.201165 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.202852 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.203079 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.206172 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.216037 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4ncp\" (UniqueName: \"kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp\") pod \"ceilometer-0\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.274355 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.812218 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.882964 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerStarted","Data":"6b451304cf169c1e26e660c2ee6383699253203c502a8e32edeac73054db3ed6"} Nov 24 08:35:16 crc kubenswrapper[4831]: I1124 08:35:16.905076 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c303ee-4bdf-4d13-9dc7-26c78fba7ce6" path="/var/lib/kubelet/pods/18c303ee-4bdf-4d13-9dc7-26c78fba7ce6/volumes" Nov 24 08:35:17 crc kubenswrapper[4831]: I1124 08:35:17.086940 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:17 crc kubenswrapper[4831]: I1124 08:35:17.934077 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerStarted","Data":"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375"} Nov 24 08:35:18 crc kubenswrapper[4831]: E1124 08:35:18.361202 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod503f7d3b_71dc_4fc1_b2be_858dd654db32.slice/crio-conmon-9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.550976 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.650627 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle\") pod \"503f7d3b-71dc-4fc1-b2be-858dd654db32\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.651006 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data\") pod \"503f7d3b-71dc-4fc1-b2be-858dd654db32\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.651049 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs\") pod \"503f7d3b-71dc-4fc1-b2be-858dd654db32\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.651080 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glmch\" (UniqueName: \"kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch\") pod \"503f7d3b-71dc-4fc1-b2be-858dd654db32\" (UID: \"503f7d3b-71dc-4fc1-b2be-858dd654db32\") " Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.659669 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs" (OuterVolumeSpecName: "logs") pod "503f7d3b-71dc-4fc1-b2be-858dd654db32" (UID: "503f7d3b-71dc-4fc1-b2be-858dd654db32"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.665001 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch" (OuterVolumeSpecName: "kube-api-access-glmch") pod "503f7d3b-71dc-4fc1-b2be-858dd654db32" (UID: "503f7d3b-71dc-4fc1-b2be-858dd654db32"). InnerVolumeSpecName "kube-api-access-glmch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.719556 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data" (OuterVolumeSpecName: "config-data") pod "503f7d3b-71dc-4fc1-b2be-858dd654db32" (UID: "503f7d3b-71dc-4fc1-b2be-858dd654db32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.741746 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "503f7d3b-71dc-4fc1-b2be-858dd654db32" (UID: "503f7d3b-71dc-4fc1-b2be-858dd654db32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.753244 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.753279 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/503f7d3b-71dc-4fc1-b2be-858dd654db32-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.753290 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/503f7d3b-71dc-4fc1-b2be-858dd654db32-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.753300 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glmch\" (UniqueName: \"kubernetes.io/projected/503f7d3b-71dc-4fc1-b2be-858dd654db32-kube-api-access-glmch\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.944629 4831 generic.go:334] "Generic (PLEG): container finished" podID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerID="9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9" exitCode=0 Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.944682 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.944693 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerDied","Data":"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9"} Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.944751 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"503f7d3b-71dc-4fc1-b2be-858dd654db32","Type":"ContainerDied","Data":"82f19cb8e4c0454dd3996cb52f8c09cabd237a9f89e50bb9f217e4ab2e164e71"} Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.944788 4831 scope.go:117] "RemoveContainer" containerID="9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.953357 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerStarted","Data":"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d"} Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.953534 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerStarted","Data":"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0"} Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.973059 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.977418 4831 scope.go:117] "RemoveContainer" containerID="d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be" Nov 24 08:35:18 crc kubenswrapper[4831]: I1124 08:35:18.983450 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.004225 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:19 crc kubenswrapper[4831]: E1124 08:35:19.004640 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-log" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.004655 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-log" Nov 24 08:35:19 crc kubenswrapper[4831]: E1124 08:35:19.004666 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-api" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.004672 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-api" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.004870 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-log" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.004886 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" containerName="nova-api-api" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.005818 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.010654 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.011262 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.011543 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.019727 4831 scope.go:117] "RemoveContainer" containerID="9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.020223 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:19 crc kubenswrapper[4831]: E1124 08:35:19.020602 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9\": container with ID starting with 9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9 not found: ID does not exist" containerID="9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.020700 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9"} err="failed to get container status \"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9\": rpc error: code = NotFound desc = could not find container \"9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9\": container with ID starting with 9974b65022fd44b75b09f261a5fd74ba8eca00336138366112f731c0f70b0bf9 not found: ID does not exist" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.020776 4831 scope.go:117] "RemoveContainer" containerID="d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be" Nov 24 08:35:19 crc kubenswrapper[4831]: E1124 08:35:19.021238 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be\": container with ID starting with d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be not found: ID does not exist" containerID="d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.021277 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be"} err="failed to get container status \"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be\": rpc error: code = NotFound desc = could not find container \"d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be\": container with ID starting with d232f4644a5114a783c6a4beb02534005eef517e646d8e16e6798e9fe403f6be not found: ID does not exist" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159269 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltdqp\" (UniqueName: \"kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159318 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159383 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159401 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159423 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.159454 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261035 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltdqp\" (UniqueName: \"kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261410 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261463 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261484 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261503 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.261533 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.262319 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.270098 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.270143 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.271999 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.288821 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.294884 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltdqp\" (UniqueName: \"kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp\") pod \"nova-api-0\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.322045 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.813830 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:19 crc kubenswrapper[4831]: I1124 08:35:19.964290 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerStarted","Data":"b50e069e3475c5e06bf0a53ecdec1828e77c42be8f80d69802a7e3844dc2b2b3"} Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.146088 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.168975 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.169024 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.177723 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.938658 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="503f7d3b-71dc-4fc1-b2be-858dd654db32" path="/var/lib/kubelet/pods/503f7d3b-71dc-4fc1-b2be-858dd654db32/volumes" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.980357 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerStarted","Data":"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59"} Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.980628 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-central-agent" containerID="cri-o://71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375" gracePeriod=30 Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.982457 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="sg-core" containerID="cri-o://2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d" gracePeriod=30 Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.982490 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-notification-agent" containerID="cri-o://8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0" gracePeriod=30 Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.982457 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="proxy-httpd" containerID="cri-o://7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59" gracePeriod=30 Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.982994 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.991411 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerStarted","Data":"df2d2ea91fefd6303d836b9077d1b60b8810b0000f7cf6c07a59abc89605cdc3"} Nov 24 08:35:20 crc kubenswrapper[4831]: I1124 08:35:20.991454 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerStarted","Data":"09c733586dead3a61e6050d2aa4e8ef21bb3acf8bf26797ed2866e351f4dc4f8"} Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.010846 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.170625639 podStartE2EDuration="6.010818446s" podCreationTimestamp="2025-11-24 08:35:15 +0000 UTC" firstStartedPulling="2025-11-24 08:35:16.815207684 +0000 UTC m=+1190.690352837" lastFinishedPulling="2025-11-24 08:35:20.655400491 +0000 UTC m=+1194.530545644" observedRunningTime="2025-11-24 08:35:21.004896687 +0000 UTC m=+1194.880041850" watchObservedRunningTime="2025-11-24 08:35:21.010818446 +0000 UTC m=+1194.885963599" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.032222 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.032205066 podStartE2EDuration="3.032205066s" podCreationTimestamp="2025-11-24 08:35:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:21.025741432 +0000 UTC m=+1194.900886595" watchObservedRunningTime="2025-11-24 08:35:21.032205066 +0000 UTC m=+1194.907350219" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.182589 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.182653 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.299756 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.471500 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-22ldq"] Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.472901 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.479452 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-22ldq"] Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.481250 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.481479 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.604565 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.604829 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.605024 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.605456 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmm5r\" (UniqueName: \"kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.605532 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.670014 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.670255 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="dnsmasq-dns" containerID="cri-o://d9403092dffd26cc22c5dbdc758b999a3397debbf39689e5349ec3428fa6fd6c" gracePeriod=10 Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.707458 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.707589 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmm5r\" (UniqueName: \"kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.707613 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.707666 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.715886 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.719344 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.727098 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.739964 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmm5r\" (UniqueName: \"kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r\") pod \"nova-cell1-cell-mapping-22ldq\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:21 crc kubenswrapper[4831]: I1124 08:35:21.787458 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.033809 4831 generic.go:334] "Generic (PLEG): container finished" podID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerID="2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d" exitCode=2 Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.034130 4831 generic.go:334] "Generic (PLEG): container finished" podID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerID="8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0" exitCode=0 Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.033963 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerDied","Data":"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d"} Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.034181 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerDied","Data":"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0"} Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.062857 4831 generic.go:334] "Generic (PLEG): container finished" podID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerID="d9403092dffd26cc22c5dbdc758b999a3397debbf39689e5349ec3428fa6fd6c" exitCode=0 Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.062959 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" event={"ID":"5022f8f0-893d-4e8a-8fb8-b9967ed943d8","Type":"ContainerDied","Data":"d9403092dffd26cc22c5dbdc758b999a3397debbf39689e5349ec3428fa6fd6c"} Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.401887 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-22ldq"] Nov 24 08:35:22 crc kubenswrapper[4831]: W1124 08:35:22.416462 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2497432_87c5_4ec8_8bd0_5a27e60e6c45.slice/crio-ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9 WatchSource:0}: Error finding container ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9: Status 404 returned error can't find the container with id ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9 Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.517997 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.632755 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config\") pod \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.633262 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb\") pod \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.633346 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb\") pod \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.633401 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz7ws\" (UniqueName: \"kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws\") pod \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.633898 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc\") pod \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\" (UID: \"5022f8f0-893d-4e8a-8fb8-b9967ed943d8\") " Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.636570 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws" (OuterVolumeSpecName: "kube-api-access-cz7ws") pod "5022f8f0-893d-4e8a-8fb8-b9967ed943d8" (UID: "5022f8f0-893d-4e8a-8fb8-b9967ed943d8"). InnerVolumeSpecName "kube-api-access-cz7ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.689955 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5022f8f0-893d-4e8a-8fb8-b9967ed943d8" (UID: "5022f8f0-893d-4e8a-8fb8-b9967ed943d8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.699006 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5022f8f0-893d-4e8a-8fb8-b9967ed943d8" (UID: "5022f8f0-893d-4e8a-8fb8-b9967ed943d8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.701778 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config" (OuterVolumeSpecName: "config") pod "5022f8f0-893d-4e8a-8fb8-b9967ed943d8" (UID: "5022f8f0-893d-4e8a-8fb8-b9967ed943d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.717332 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5022f8f0-893d-4e8a-8fb8-b9967ed943d8" (UID: "5022f8f0-893d-4e8a-8fb8-b9967ed943d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.736025 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.736249 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.736315 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.736397 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz7ws\" (UniqueName: \"kubernetes.io/projected/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-kube-api-access-cz7ws\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:22 crc kubenswrapper[4831]: I1124 08:35:22.736465 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5022f8f0-893d-4e8a-8fb8-b9967ed943d8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.077826 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" event={"ID":"5022f8f0-893d-4e8a-8fb8-b9967ed943d8","Type":"ContainerDied","Data":"b0ecb45204872b7203e1fc16962a5e597f2e193aed01cc58c543e4eac5daf2ba"} Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.078224 4831 scope.go:117] "RemoveContainer" containerID="d9403092dffd26cc22c5dbdc758b999a3397debbf39689e5349ec3428fa6fd6c" Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.077950 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-26cdj" Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.080107 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-22ldq" event={"ID":"c2497432-87c5-4ec8-8bd0-5a27e60e6c45","Type":"ContainerStarted","Data":"5df5ec28960fdd9ac8045f4be2310b118b924023266fd3addf77d19160cfab20"} Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.080141 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-22ldq" event={"ID":"c2497432-87c5-4ec8-8bd0-5a27e60e6c45","Type":"ContainerStarted","Data":"ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9"} Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.098597 4831 scope.go:117] "RemoveContainer" containerID="a38711d3e958f5678dbac631e43d5558b0c423f716aadc01030c8635dee9fa49" Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.111499 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-22ldq" podStartSLOduration=2.111480308 podStartE2EDuration="2.111480308s" podCreationTimestamp="2025-11-24 08:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:23.102931015 +0000 UTC m=+1196.978076168" watchObservedRunningTime="2025-11-24 08:35:23.111480308 +0000 UTC m=+1196.986625461" Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.127611 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:35:23 crc kubenswrapper[4831]: I1124 08:35:23.138635 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-26cdj"] Nov 24 08:35:24 crc kubenswrapper[4831]: I1124 08:35:24.099585 4831 generic.go:334] "Generic (PLEG): container finished" podID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerID="71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375" exitCode=0 Nov 24 08:35:24 crc kubenswrapper[4831]: I1124 08:35:24.100034 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerDied","Data":"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375"} Nov 24 08:35:24 crc kubenswrapper[4831]: I1124 08:35:24.914789 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" path="/var/lib/kubelet/pods/5022f8f0-893d-4e8a-8fb8-b9967ed943d8/volumes" Nov 24 08:35:28 crc kubenswrapper[4831]: I1124 08:35:28.139291 4831 generic.go:334] "Generic (PLEG): container finished" podID="c2497432-87c5-4ec8-8bd0-5a27e60e6c45" containerID="5df5ec28960fdd9ac8045f4be2310b118b924023266fd3addf77d19160cfab20" exitCode=0 Nov 24 08:35:28 crc kubenswrapper[4831]: I1124 08:35:28.139374 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-22ldq" event={"ID":"c2497432-87c5-4ec8-8bd0-5a27e60e6c45","Type":"ContainerDied","Data":"5df5ec28960fdd9ac8045f4be2310b118b924023266fd3addf77d19160cfab20"} Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.322763 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.323099 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.559213 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.669881 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data\") pod \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.669936 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle\") pod \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.669995 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmm5r\" (UniqueName: \"kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r\") pod \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.670136 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts\") pod \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\" (UID: \"c2497432-87c5-4ec8-8bd0-5a27e60e6c45\") " Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.693306 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r" (OuterVolumeSpecName: "kube-api-access-hmm5r") pod "c2497432-87c5-4ec8-8bd0-5a27e60e6c45" (UID: "c2497432-87c5-4ec8-8bd0-5a27e60e6c45"). InnerVolumeSpecName "kube-api-access-hmm5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.693461 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts" (OuterVolumeSpecName: "scripts") pod "c2497432-87c5-4ec8-8bd0-5a27e60e6c45" (UID: "c2497432-87c5-4ec8-8bd0-5a27e60e6c45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.701043 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data" (OuterVolumeSpecName: "config-data") pod "c2497432-87c5-4ec8-8bd0-5a27e60e6c45" (UID: "c2497432-87c5-4ec8-8bd0-5a27e60e6c45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.706128 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2497432-87c5-4ec8-8bd0-5a27e60e6c45" (UID: "c2497432-87c5-4ec8-8bd0-5a27e60e6c45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.772346 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.772388 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.772404 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmm5r\" (UniqueName: \"kubernetes.io/projected/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-kube-api-access-hmm5r\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:29 crc kubenswrapper[4831]: I1124 08:35:29.772417 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2497432-87c5-4ec8-8bd0-5a27e60e6c45-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.162984 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-22ldq" event={"ID":"c2497432-87c5-4ec8-8bd0-5a27e60e6c45","Type":"ContainerDied","Data":"ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9"} Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.163396 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ada8a36642895bc010b8f3c3dfe58bcd6e28f24ae278d33c44fffa54e93de7a9" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.163212 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-22ldq" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.186335 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.210014 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.210694 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.334831 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.334909 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.185:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.350458 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.351394 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-log" containerID="cri-o://09c733586dead3a61e6050d2aa4e8ef21bb3acf8bf26797ed2866e351f4dc4f8" gracePeriod=30 Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.351793 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-api" containerID="cri-o://df2d2ea91fefd6303d836b9077d1b60b8810b0000f7cf6c07a59abc89605cdc3" gracePeriod=30 Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.375126 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.375354 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerName="nova-scheduler-scheduler" containerID="cri-o://0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" gracePeriod=30 Nov 24 08:35:30 crc kubenswrapper[4831]: I1124 08:35:30.417943 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:31 crc kubenswrapper[4831]: I1124 08:35:31.172156 4831 generic.go:334] "Generic (PLEG): container finished" podID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerID="09c733586dead3a61e6050d2aa4e8ef21bb3acf8bf26797ed2866e351f4dc4f8" exitCode=143 Nov 24 08:35:31 crc kubenswrapper[4831]: I1124 08:35:31.174449 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerDied","Data":"09c733586dead3a61e6050d2aa4e8ef21bb3acf8bf26797ed2866e351f4dc4f8"} Nov 24 08:35:31 crc kubenswrapper[4831]: I1124 08:35:31.181632 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:35:31 crc kubenswrapper[4831]: E1124 08:35:31.845524 4831 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:35:31 crc kubenswrapper[4831]: E1124 08:35:31.847033 4831 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:35:31 crc kubenswrapper[4831]: E1124 08:35:31.848617 4831 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 24 08:35:31 crc kubenswrapper[4831]: E1124 08:35:31.848686 4831 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerName="nova-scheduler-scheduler" Nov 24 08:35:32 crc kubenswrapper[4831]: I1124 08:35:32.181506 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" containerID="cri-o://7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7" gracePeriod=30 Nov 24 08:35:32 crc kubenswrapper[4831]: I1124 08:35:32.181905 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" containerID="cri-o://ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a" gracePeriod=30 Nov 24 08:35:33 crc kubenswrapper[4831]: I1124 08:35:33.190603 4831 generic.go:334] "Generic (PLEG): container finished" podID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerID="7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7" exitCode=143 Nov 24 08:35:33 crc kubenswrapper[4831]: I1124 08:35:33.190647 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerDied","Data":"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7"} Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.255267 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": read tcp 10.217.0.2:54650->10.217.0.182:8775: read: connection reset by peer" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.256079 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.182:8775/\": read tcp 10.217.0.2:54652->10.217.0.182:8775: read: connection reset by peer" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.792925 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.898460 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data\") pod \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.898501 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle\") pod \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.898542 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs\") pod \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.898682 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs\") pod \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.898723 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgvfd\" (UniqueName: \"kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd\") pod \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\" (UID: \"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04\") " Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.901176 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs" (OuterVolumeSpecName: "logs") pod "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" (UID: "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.905534 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd" (OuterVolumeSpecName: "kube-api-access-fgvfd") pod "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" (UID: "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04"). InnerVolumeSpecName "kube-api-access-fgvfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.927507 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" (UID: "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.946194 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data" (OuterVolumeSpecName: "config-data") pod "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" (UID: "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:35 crc kubenswrapper[4831]: I1124 08:35:35.985636 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" (UID: "6b8c6a67-2bd5-403d-8c0a-6adb179b0d04"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.000765 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.000789 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.000801 4831 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.000810 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.000819 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgvfd\" (UniqueName: \"kubernetes.io/projected/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04-kube-api-access-fgvfd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.165113 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.217240 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data\") pod \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.217370 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvrld\" (UniqueName: \"kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld\") pod \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.217571 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle\") pod \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\" (UID: \"b518f57f-cc58-4b1e-be5e-74bc46526a4f\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.225389 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld" (OuterVolumeSpecName: "kube-api-access-fvrld") pod "b518f57f-cc58-4b1e-be5e-74bc46526a4f" (UID: "b518f57f-cc58-4b1e-be5e-74bc46526a4f"). InnerVolumeSpecName "kube-api-access-fvrld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.226583 4831 generic.go:334] "Generic (PLEG): container finished" podID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" exitCode=0 Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.226647 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b518f57f-cc58-4b1e-be5e-74bc46526a4f","Type":"ContainerDied","Data":"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82"} Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.226673 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b518f57f-cc58-4b1e-be5e-74bc46526a4f","Type":"ContainerDied","Data":"5d1973761458280e728b9e6dc145e3593d91e78f91ec4c7e3b98027711a490a8"} Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.226692 4831 scope.go:117] "RemoveContainer" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.226899 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.231877 4831 generic.go:334] "Generic (PLEG): container finished" podID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerID="df2d2ea91fefd6303d836b9077d1b60b8810b0000f7cf6c07a59abc89605cdc3" exitCode=0 Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.231935 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerDied","Data":"df2d2ea91fefd6303d836b9077d1b60b8810b0000f7cf6c07a59abc89605cdc3"} Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.233873 4831 generic.go:334] "Generic (PLEG): container finished" podID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerID="ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a" exitCode=0 Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.233898 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerDied","Data":"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a"} Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.234112 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b8c6a67-2bd5-403d-8c0a-6adb179b0d04","Type":"ContainerDied","Data":"da7f53de936bc12546b7e100116fffa99ec9e4e0daaf7ff219ae2384f22beeb1"} Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.234173 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.257754 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data" (OuterVolumeSpecName: "config-data") pod "b518f57f-cc58-4b1e-be5e-74bc46526a4f" (UID: "b518f57f-cc58-4b1e-be5e-74bc46526a4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.264430 4831 scope.go:117] "RemoveContainer" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.267157 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82\": container with ID starting with 0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82 not found: ID does not exist" containerID="0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.267212 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82"} err="failed to get container status \"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82\": rpc error: code = NotFound desc = could not find container \"0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82\": container with ID starting with 0f657961fe6c1ac1386342d17c5fcd13be787efe7881180ff491d0a2fffe6c82 not found: ID does not exist" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.267249 4831 scope.go:117] "RemoveContainer" containerID="ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.279209 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b518f57f-cc58-4b1e-be5e-74bc46526a4f" (UID: "b518f57f-cc58-4b1e-be5e-74bc46526a4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.283447 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.289490 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.294192 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316286 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316723 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerName="nova-scheduler-scheduler" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316752 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerName="nova-scheduler-scheduler" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316767 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316773 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316783 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-api" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316789 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-api" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316809 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="dnsmasq-dns" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316815 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="dnsmasq-dns" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316830 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="init" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316837 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="init" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316847 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2497432-87c5-4ec8-8bd0-5a27e60e6c45" containerName="nova-manage" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316853 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2497432-87c5-4ec8-8bd0-5a27e60e6c45" containerName="nova-manage" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316862 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-log" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316868 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-log" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.316879 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.316884 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317043 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-log" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317052 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-log" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317062 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" containerName="nova-metadata-metadata" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317076 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" containerName="nova-scheduler-scheduler" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317088 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2497432-87c5-4ec8-8bd0-5a27e60e6c45" containerName="nova-manage" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317096 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5022f8f0-893d-4e8a-8fb8-b9967ed943d8" containerName="dnsmasq-dns" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.317104 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" containerName="nova-api-api" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.318030 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.323071 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.330758 4831 scope.go:117] "RemoveContainer" containerID="7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.331183 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.332766 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.332797 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvrld\" (UniqueName: \"kubernetes.io/projected/b518f57f-cc58-4b1e-be5e-74bc46526a4f-kube-api-access-fvrld\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.332808 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b518f57f-cc58-4b1e-be5e-74bc46526a4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.340115 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.401059 4831 scope.go:117] "RemoveContainer" containerID="ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.401721 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a\": container with ID starting with ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a not found: ID does not exist" containerID="ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.401766 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a"} err="failed to get container status \"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a\": rpc error: code = NotFound desc = could not find container \"ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a\": container with ID starting with ac44a6638706663fcf0d1586c6e93978b5fa05aa7514218a95a5d3dc056e3e1a not found: ID does not exist" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.401795 4831 scope.go:117] "RemoveContainer" containerID="7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7" Nov 24 08:35:36 crc kubenswrapper[4831]: E1124 08:35:36.402091 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7\": container with ID starting with 7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7 not found: ID does not exist" containerID="7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.402116 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7"} err="failed to get container status \"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7\": rpc error: code = NotFound desc = could not find container \"7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7\": container with ID starting with 7cfaeaa14b13ab0952f72533530e0bf9bfc9f80836ee688e6a89dd98b981b9d7 not found: ID does not exist" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434477 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434596 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434655 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434711 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434739 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltdqp\" (UniqueName: \"kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.434813 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs\") pod \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\" (UID: \"3078bad4-b574-44ed-b3a1-ee36e4b02e40\") " Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435161 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435229 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvllb\" (UniqueName: \"kubernetes.io/projected/86f061df-2ac1-4323-b6ce-b6c710d98793-kube-api-access-xvllb\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435345 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435374 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-config-data\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435422 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f061df-2ac1-4323-b6ce-b6c710d98793-logs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.435905 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs" (OuterVolumeSpecName: "logs") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.440306 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp" (OuterVolumeSpecName: "kube-api-access-ltdqp") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "kube-api-access-ltdqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.477000 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.477524 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.481161 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data" (OuterVolumeSpecName: "config-data") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.482978 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3078bad4-b574-44ed-b3a1-ee36e4b02e40" (UID: "3078bad4-b574-44ed-b3a1-ee36e4b02e40"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537326 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvllb\" (UniqueName: \"kubernetes.io/projected/86f061df-2ac1-4323-b6ce-b6c710d98793-kube-api-access-xvllb\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537429 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537456 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-config-data\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537494 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f061df-2ac1-4323-b6ce-b6c710d98793-logs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537529 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537586 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537596 4831 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3078bad4-b574-44ed-b3a1-ee36e4b02e40-logs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537605 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltdqp\" (UniqueName: \"kubernetes.io/projected/3078bad4-b574-44ed-b3a1-ee36e4b02e40-kube-api-access-ltdqp\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537614 4831 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537622 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.537630 4831 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3078bad4-b574-44ed-b3a1-ee36e4b02e40-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.538558 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f061df-2ac1-4323-b6ce-b6c710d98793-logs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.540556 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.542145 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-config-data\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.550802 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f061df-2ac1-4323-b6ce-b6c710d98793-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.554331 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvllb\" (UniqueName: \"kubernetes.io/projected/86f061df-2ac1-4323-b6ce-b6c710d98793-kube-api-access-xvllb\") pod \"nova-metadata-0\" (UID: \"86f061df-2ac1-4323-b6ce-b6c710d98793\") " pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.629872 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.650216 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.669593 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.670867 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.674965 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.677346 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.703461 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.741448 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9lcl\" (UniqueName: \"kubernetes.io/projected/6df77d90-f9a7-4653-83e0-d465588ed50e-kube-api-access-f9lcl\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.741661 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.741860 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-config-data\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.843827 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.843901 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-config-data\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.843940 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9lcl\" (UniqueName: \"kubernetes.io/projected/6df77d90-f9a7-4653-83e0-d465588ed50e-kube-api-access-f9lcl\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.849226 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.852695 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df77d90-f9a7-4653-83e0-d465588ed50e-config-data\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.916130 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9lcl\" (UniqueName: \"kubernetes.io/projected/6df77d90-f9a7-4653-83e0-d465588ed50e-kube-api-access-f9lcl\") pod \"nova-scheduler-0\" (UID: \"6df77d90-f9a7-4653-83e0-d465588ed50e\") " pod="openstack/nova-scheduler-0" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.923239 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b8c6a67-2bd5-403d-8c0a-6adb179b0d04" path="/var/lib/kubelet/pods/6b8c6a67-2bd5-403d-8c0a-6adb179b0d04/volumes" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.923849 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b518f57f-cc58-4b1e-be5e-74bc46526a4f" path="/var/lib/kubelet/pods/b518f57f-cc58-4b1e-be5e-74bc46526a4f/volumes" Nov 24 08:35:36 crc kubenswrapper[4831]: I1124 08:35:36.990091 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.213682 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: W1124 08:35:37.223261 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86f061df_2ac1_4323_b6ce_b6c710d98793.slice/crio-7e31292c813d3e296f4a7fe1c1b923e72e21c967c158e5282277f83c92fe1853 WatchSource:0}: Error finding container 7e31292c813d3e296f4a7fe1c1b923e72e21c967c158e5282277f83c92fe1853: Status 404 returned error can't find the container with id 7e31292c813d3e296f4a7fe1c1b923e72e21c967c158e5282277f83c92fe1853 Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.248614 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86f061df-2ac1-4323-b6ce-b6c710d98793","Type":"ContainerStarted","Data":"7e31292c813d3e296f4a7fe1c1b923e72e21c967c158e5282277f83c92fe1853"} Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.261918 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3078bad4-b574-44ed-b3a1-ee36e4b02e40","Type":"ContainerDied","Data":"b50e069e3475c5e06bf0a53ecdec1828e77c42be8f80d69802a7e3844dc2b2b3"} Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.261980 4831 scope.go:117] "RemoveContainer" containerID="df2d2ea91fefd6303d836b9077d1b60b8810b0000f7cf6c07a59abc89605cdc3" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.262501 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.292050 4831 scope.go:117] "RemoveContainer" containerID="09c733586dead3a61e6050d2aa4e8ef21bb3acf8bf26797ed2866e351f4dc4f8" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.308263 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.319713 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.326881 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.328214 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.335822 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.335953 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.335985 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.373347 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470478 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470526 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de07ebb9-282e-4dd6-9a4f-8267816e827d-logs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470579 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-public-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470609 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470680 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-config-data\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.470724 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtllt\" (UniqueName: \"kubernetes.io/projected/de07ebb9-282e-4dd6-9a4f-8267816e827d-kube-api-access-qtllt\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.568626 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.571758 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.575126 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-config-data\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.575538 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtllt\" (UniqueName: \"kubernetes.io/projected/de07ebb9-282e-4dd6-9a4f-8267816e827d-kube-api-access-qtllt\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.575723 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.575827 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de07ebb9-282e-4dd6-9a4f-8267816e827d-logs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.576128 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-public-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.577148 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de07ebb9-282e-4dd6-9a4f-8267816e827d-logs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.582219 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.585526 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-public-tls-certs\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.625142 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtllt\" (UniqueName: \"kubernetes.io/projected/de07ebb9-282e-4dd6-9a4f-8267816e827d-kube-api-access-qtllt\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.635531 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-config-data\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.635989 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de07ebb9-282e-4dd6-9a4f-8267816e827d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de07ebb9-282e-4dd6-9a4f-8267816e827d\") " pod="openstack/nova-api-0" Nov 24 08:35:37 crc kubenswrapper[4831]: I1124 08:35:37.744699 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 24 08:35:38 crc kubenswrapper[4831]: W1124 08:35:38.216763 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde07ebb9_282e_4dd6_9a4f_8267816e827d.slice/crio-d4f18743520d5c59ac549bb12db84e9c8cee4104e5e77461ab5fde7c1106a540 WatchSource:0}: Error finding container d4f18743520d5c59ac549bb12db84e9c8cee4104e5e77461ab5fde7c1106a540: Status 404 returned error can't find the container with id d4f18743520d5c59ac549bb12db84e9c8cee4104e5e77461ab5fde7c1106a540 Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.219529 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.278465 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de07ebb9-282e-4dd6-9a4f-8267816e827d","Type":"ContainerStarted","Data":"d4f18743520d5c59ac549bb12db84e9c8cee4104e5e77461ab5fde7c1106a540"} Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.280272 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86f061df-2ac1-4323-b6ce-b6c710d98793","Type":"ContainerStarted","Data":"803cbcaa10aa8077db5918d8bfd8f2e999735186bb32a578f914526bdf909e77"} Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.280410 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"86f061df-2ac1-4323-b6ce-b6c710d98793","Type":"ContainerStarted","Data":"a0d38ad677ad2035cc53126e35811a8a4e88bf1e8b5b0c04f65e7ce282990fbf"} Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.282024 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6df77d90-f9a7-4653-83e0-d465588ed50e","Type":"ContainerStarted","Data":"8734418796f89e2f3252fe954fbb9d82492185661cac0ffc5baad95f49f2938c"} Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.282060 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6df77d90-f9a7-4653-83e0-d465588ed50e","Type":"ContainerStarted","Data":"75ff531c4ac0a853586e635682128fbf1a983c25d43cff2190e81bc9862d81c1"} Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.316565 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.316545798 podStartE2EDuration="2.316545798s" podCreationTimestamp="2025-11-24 08:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:38.297733411 +0000 UTC m=+1212.172878594" watchObservedRunningTime="2025-11-24 08:35:38.316545798 +0000 UTC m=+1212.191690961" Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.334972 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.334955203 podStartE2EDuration="2.334955203s" podCreationTimestamp="2025-11-24 08:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:38.328516539 +0000 UTC m=+1212.203661702" watchObservedRunningTime="2025-11-24 08:35:38.334955203 +0000 UTC m=+1212.210100356" Nov 24 08:35:38 crc kubenswrapper[4831]: I1124 08:35:38.908760 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3078bad4-b574-44ed-b3a1-ee36e4b02e40" path="/var/lib/kubelet/pods/3078bad4-b574-44ed-b3a1-ee36e4b02e40/volumes" Nov 24 08:35:39 crc kubenswrapper[4831]: I1124 08:35:39.303639 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de07ebb9-282e-4dd6-9a4f-8267816e827d","Type":"ContainerStarted","Data":"17731280e6e0aabb8c26bf651eb267af471eb593b75384e697452cb698a1a85c"} Nov 24 08:35:39 crc kubenswrapper[4831]: I1124 08:35:39.304078 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de07ebb9-282e-4dd6-9a4f-8267816e827d","Type":"ContainerStarted","Data":"63867c0e3d11b6d64369f3ea27863bee7bc425bc6bf492bb3093320f03a1484b"} Nov 24 08:35:39 crc kubenswrapper[4831]: I1124 08:35:39.327486 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.327464596 podStartE2EDuration="2.327464596s" podCreationTimestamp="2025-11-24 08:35:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:35:39.324309316 +0000 UTC m=+1213.199454509" watchObservedRunningTime="2025-11-24 08:35:39.327464596 +0000 UTC m=+1213.202609749" Nov 24 08:35:41 crc kubenswrapper[4831]: I1124 08:35:41.704246 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:35:41 crc kubenswrapper[4831]: I1124 08:35:41.704585 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 24 08:35:41 crc kubenswrapper[4831]: I1124 08:35:41.991331 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 24 08:35:46 crc kubenswrapper[4831]: I1124 08:35:46.287587 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 24 08:35:46 crc kubenswrapper[4831]: I1124 08:35:46.707698 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:35:46 crc kubenswrapper[4831]: I1124 08:35:46.707781 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 24 08:35:46 crc kubenswrapper[4831]: I1124 08:35:46.990947 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.018613 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.398949 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.716572 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="86f061df-2ac1-4323-b6ce-b6c710d98793" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.724477 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="86f061df-2ac1-4323-b6ce-b6c710d98793" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.745969 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:47 crc kubenswrapper[4831]: I1124 08:35:47.746019 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 24 08:35:48 crc kubenswrapper[4831]: I1124 08:35:48.795507 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de07ebb9-282e-4dd6-9a4f-8267816e827d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.189:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:48 crc kubenswrapper[4831]: I1124 08:35:48.795507 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de07ebb9-282e-4dd6-9a4f-8267816e827d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.189:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.370007 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.417779 4831 generic.go:334] "Generic (PLEG): container finished" podID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerID="7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59" exitCode=137 Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.417843 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerDied","Data":"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59"} Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.417865 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.417891 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1","Type":"ContainerDied","Data":"6b451304cf169c1e26e660c2ee6383699253203c502a8e32edeac73054db3ed6"} Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.417931 4831 scope.go:117] "RemoveContainer" containerID="7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.439840 4831 scope.go:117] "RemoveContainer" containerID="2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.463237 4831 scope.go:117] "RemoveContainer" containerID="8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471044 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471134 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471178 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471224 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4ncp\" (UniqueName: \"kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471286 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471335 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471499 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.471528 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs\") pod \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\" (UID: \"99be40d2-ff4e-4fae-a5f1-e116ae4f66b1\") " Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.472821 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.475857 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.477551 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts" (OuterVolumeSpecName: "scripts") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.480636 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp" (OuterVolumeSpecName: "kube-api-access-z4ncp") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "kube-api-access-z4ncp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.484827 4831 scope.go:117] "RemoveContainer" containerID="71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.503989 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.538250 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.553292 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573077 4831 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573106 4831 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573114 4831 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573123 4831 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-scripts\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573132 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4ncp\" (UniqueName: \"kubernetes.io/projected/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-kube-api-access-z4ncp\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573141 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.573150 4831 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.580698 4831 scope.go:117] "RemoveContainer" containerID="7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.581157 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59\": container with ID starting with 7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59 not found: ID does not exist" containerID="7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.581208 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59"} err="failed to get container status \"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59\": rpc error: code = NotFound desc = could not find container \"7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59\": container with ID starting with 7cba243d182a189a1e7317c863fc069dc5e91c561dc085d554199a0e727e9a59 not found: ID does not exist" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.581234 4831 scope.go:117] "RemoveContainer" containerID="2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.581734 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d\": container with ID starting with 2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d not found: ID does not exist" containerID="2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.581758 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d"} err="failed to get container status \"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d\": rpc error: code = NotFound desc = could not find container \"2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d\": container with ID starting with 2f6598d4dfe554c08e16c4c7a2d35fb32cdb095d8b707c41a1450bffcc7c8e6d not found: ID does not exist" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.581771 4831 scope.go:117] "RemoveContainer" containerID="8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.582797 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0\": container with ID starting with 8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0 not found: ID does not exist" containerID="8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.582859 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0"} err="failed to get container status \"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0\": rpc error: code = NotFound desc = could not find container \"8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0\": container with ID starting with 8fe6ef14402e36b60436af0a837a7e21c099f1801063fcd40ec3a9cd08463fc0 not found: ID does not exist" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.582891 4831 scope.go:117] "RemoveContainer" containerID="71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.583354 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375\": container with ID starting with 71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375 not found: ID does not exist" containerID="71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.583384 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375"} err="failed to get container status \"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375\": rpc error: code = NotFound desc = could not find container \"71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375\": container with ID starting with 71e4b339e8870e887aac8e6469acb3da107e6e4376f67fa2925979dc9edfc375 not found: ID does not exist" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.592240 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data" (OuterVolumeSpecName: "config-data") pod "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" (UID: "99be40d2-ff4e-4fae-a5f1-e116ae4f66b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.674865 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.753637 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.774519 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.781878 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.782283 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-central-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782307 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-central-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.782353 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="sg-core" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782361 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="sg-core" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.782371 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-notification-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782377 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-notification-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: E1124 08:35:51.782399 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="proxy-httpd" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782405 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="proxy-httpd" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782554 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-central-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782577 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="proxy-httpd" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782587 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="sg-core" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.782597 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" containerName="ceilometer-notification-agent" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.784234 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.804277 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.804453 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.804725 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.810087 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878539 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878582 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-config-data\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878607 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-run-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878633 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6xmv\" (UniqueName: \"kubernetes.io/projected/21235403-7ca3-4341-9f1e-661e7e833b32-kube-api-access-m6xmv\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878653 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878683 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-scripts\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878885 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-log-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.878907 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980327 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980456 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-config-data\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980495 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-run-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980533 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6xmv\" (UniqueName: \"kubernetes.io/projected/21235403-7ca3-4341-9f1e-661e7e833b32-kube-api-access-m6xmv\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980565 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980665 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-scripts\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980801 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-log-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.980861 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.982896 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-run-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.983172 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/21235403-7ca3-4341-9f1e-661e7e833b32-log-httpd\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.985776 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.985815 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-config-data\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.987200 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.991228 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-scripts\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:51 crc kubenswrapper[4831]: I1124 08:35:51.991394 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21235403-7ca3-4341-9f1e-661e7e833b32-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:52 crc kubenswrapper[4831]: I1124 08:35:52.009565 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6xmv\" (UniqueName: \"kubernetes.io/projected/21235403-7ca3-4341-9f1e-661e7e833b32-kube-api-access-m6xmv\") pod \"ceilometer-0\" (UID: \"21235403-7ca3-4341-9f1e-661e7e833b32\") " pod="openstack/ceilometer-0" Nov 24 08:35:52 crc kubenswrapper[4831]: I1124 08:35:52.102661 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 24 08:35:52 crc kubenswrapper[4831]: I1124 08:35:52.587984 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 24 08:35:52 crc kubenswrapper[4831]: I1124 08:35:52.909273 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99be40d2-ff4e-4fae-a5f1-e116ae4f66b1" path="/var/lib/kubelet/pods/99be40d2-ff4e-4fae-a5f1-e116ae4f66b1/volumes" Nov 24 08:35:53 crc kubenswrapper[4831]: I1124 08:35:53.439111 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21235403-7ca3-4341-9f1e-661e7e833b32","Type":"ContainerStarted","Data":"c200e3f76e1ade19650e4cdc29ffe0d24154aee5a65327bce7adf3af34613835"} Nov 24 08:35:53 crc kubenswrapper[4831]: I1124 08:35:53.439636 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21235403-7ca3-4341-9f1e-661e7e833b32","Type":"ContainerStarted","Data":"dcf26f41a81b305a04b574673f80c4f32967af15dbd22b613a2c77df15573f18"} Nov 24 08:35:54 crc kubenswrapper[4831]: I1124 08:35:54.449947 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21235403-7ca3-4341-9f1e-661e7e833b32","Type":"ContainerStarted","Data":"d74f9c2b3e6b39ed5178c3e12b5a1c9f8d5b8b807b761e828d16acc67878e4ec"} Nov 24 08:35:55 crc kubenswrapper[4831]: I1124 08:35:55.463619 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21235403-7ca3-4341-9f1e-661e7e833b32","Type":"ContainerStarted","Data":"752a59675a053f4b7c3475b2c9c1f5aa5262bc080b9a88c2dbee0d868337d155"} Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.475281 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"21235403-7ca3-4341-9f1e-661e7e833b32","Type":"ContainerStarted","Data":"4ff9c1dc28a6654d6b5b3770bf8e7ff9024324a8871427136ad46fd6dec232df"} Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.476120 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.504704 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.630404189 podStartE2EDuration="5.504680352s" podCreationTimestamp="2025-11-24 08:35:51 +0000 UTC" firstStartedPulling="2025-11-24 08:35:52.599485341 +0000 UTC m=+1226.474630494" lastFinishedPulling="2025-11-24 08:35:55.473761504 +0000 UTC m=+1229.348906657" observedRunningTime="2025-11-24 08:35:56.493430941 +0000 UTC m=+1230.368576094" watchObservedRunningTime="2025-11-24 08:35:56.504680352 +0000 UTC m=+1230.379825505" Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.711482 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.714202 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 24 08:35:56 crc kubenswrapper[4831]: I1124 08:35:56.724468 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:35:57 crc kubenswrapper[4831]: I1124 08:35:57.510094 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 24 08:35:57 crc kubenswrapper[4831]: I1124 08:35:57.754735 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:35:57 crc kubenswrapper[4831]: I1124 08:35:57.755341 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:35:57 crc kubenswrapper[4831]: I1124 08:35:57.758035 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 24 08:35:57 crc kubenswrapper[4831]: I1124 08:35:57.765644 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:35:58 crc kubenswrapper[4831]: I1124 08:35:58.524013 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 24 08:35:58 crc kubenswrapper[4831]: I1124 08:35:58.539066 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 24 08:36:22 crc kubenswrapper[4831]: I1124 08:36:22.114298 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 24 08:36:28 crc kubenswrapper[4831]: I1124 08:36:28.401249 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:36:28 crc kubenswrapper[4831]: I1124 08:36:28.401959 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:36:31 crc kubenswrapper[4831]: I1124 08:36:31.196714 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:31 crc kubenswrapper[4831]: I1124 08:36:31.908169 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:36 crc kubenswrapper[4831]: I1124 08:36:36.328019 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="rabbitmq" containerID="cri-o://474b53112447ea9aedc0b761e2bae117da65c20019e0c99f838485074bc9d4d9" gracePeriod=604796 Nov 24 08:36:36 crc kubenswrapper[4831]: I1124 08:36:36.400552 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="rabbitmq" containerID="cri-o://7c831df1604165a09ea1c5798d73a1ef09c377057c5c164dd214cbd9b440ed18" gracePeriod=604795 Nov 24 08:36:41 crc kubenswrapper[4831]: I1124 08:36:41.653806 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Nov 24 08:36:41 crc kubenswrapper[4831]: I1124 08:36:41.966632 4831 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.982504 4831 generic.go:334] "Generic (PLEG): container finished" podID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerID="474b53112447ea9aedc0b761e2bae117da65c20019e0c99f838485074bc9d4d9" exitCode=0 Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.982582 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerDied","Data":"474b53112447ea9aedc0b761e2bae117da65c20019e0c99f838485074bc9d4d9"} Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.982924 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f0cc709e-e5c5-40fd-924f-aa3582689305","Type":"ContainerDied","Data":"1ce3ec7a44ca880aedd4b02baa267184e1367846c10e857c03611c30900ed1c1"} Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.982937 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce3ec7a44ca880aedd4b02baa267184e1367846c10e857c03611c30900ed1c1" Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.984836 4831 generic.go:334] "Generic (PLEG): container finished" podID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerID="7c831df1604165a09ea1c5798d73a1ef09c377057c5c164dd214cbd9b440ed18" exitCode=0 Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.984865 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerDied","Data":"7c831df1604165a09ea1c5798d73a1ef09c377057c5c164dd214cbd9b440ed18"} Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.984893 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e8b0fe50-d340-4379-8be9-4367a9c81d9f","Type":"ContainerDied","Data":"d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e"} Nov 24 08:36:42 crc kubenswrapper[4831]: I1124 08:36:42.984908 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2445b2deb744865a8399ea9258a66c8da355e1c6b4440e7a980af8bde39cd8e" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.054425 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.065616 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.068506 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.068593 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.068631 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.069972 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070073 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070586 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070633 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070716 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070778 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.070851 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.071491 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.071489 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.071524 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4npc\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc\") pod \"f0cc709e-e5c5-40fd-924f-aa3582689305\" (UID: \"f0cc709e-e5c5-40fd-924f-aa3582689305\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.072287 4831 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.072357 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.076231 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info" (OuterVolumeSpecName: "pod-info") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.076439 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.076626 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.077712 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.078528 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.079462 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc" (OuterVolumeSpecName: "kube-api-access-f4npc") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "kube-api-access-f4npc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.119805 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data" (OuterVolumeSpecName: "config-data") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.173909 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.173953 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.173974 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174012 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174046 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174079 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174117 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174179 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174218 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b24tl\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174247 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174275 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf\") pod \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\" (UID: \"e8b0fe50-d340-4379-8be9-4367a9c81d9f\") " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174694 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174709 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174720 4831 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f0cc709e-e5c5-40fd-924f-aa3582689305-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174729 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174737 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4npc\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-kube-api-access-f4npc\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174745 4831 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f0cc709e-e5c5-40fd-924f-aa3582689305-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.174765 4831 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.176468 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.176813 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.180527 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.182127 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.186713 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info" (OuterVolumeSpecName: "pod-info") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.190401 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.209883 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.242165 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl" (OuterVolumeSpecName: "kube-api-access-b24tl") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "kube-api-access-b24tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.254344 4831 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.276475 4831 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.280025 4831 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.280183 4831 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e8b0fe50-d340-4379-8be9-4367a9c81d9f-pod-info\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.280268 4831 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e8b0fe50-d340-4379-8be9-4367a9c81d9f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.282111 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.282207 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.282309 4831 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.282388 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.282454 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b24tl\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-kube-api-access-b24tl\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.278032 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf" (OuterVolumeSpecName: "server-conf") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.295967 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data" (OuterVolumeSpecName: "config-data") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.324590 4831 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.331931 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf" (OuterVolumeSpecName: "server-conf") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.346340 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f0cc709e-e5c5-40fd-924f-aa3582689305" (UID: "f0cc709e-e5c5-40fd-924f-aa3582689305"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.386593 4831 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.386628 4831 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.386641 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f0cc709e-e5c5-40fd-924f-aa3582689305-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.386651 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8b0fe50-d340-4379-8be9-4367a9c81d9f-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.386663 4831 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f0cc709e-e5c5-40fd-924f-aa3582689305-server-conf\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.429787 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e8b0fe50-d340-4379-8be9-4367a9c81d9f" (UID: "e8b0fe50-d340-4379-8be9-4367a9c81d9f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.488406 4831 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e8b0fe50-d340-4379-8be9-4367a9c81d9f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.992564 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:43 crc kubenswrapper[4831]: I1124 08:36:43.992582 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.055957 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.073407 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.085636 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.100393 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113173 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: E1124 08:36:44.113589 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="setup-container" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113612 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="setup-container" Nov 24 08:36:44 crc kubenswrapper[4831]: E1124 08:36:44.113633 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113642 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: E1124 08:36:44.113659 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="setup-container" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113667 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="setup-container" Nov 24 08:36:44 crc kubenswrapper[4831]: E1124 08:36:44.113712 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113718 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113882 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.113899 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" containerName="rabbitmq" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.114782 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.127844 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.127931 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.127847 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.128118 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.128176 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7kqpn" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.128223 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.128535 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.136809 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.138700 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.150991 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151292 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151478 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151567 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151622 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151746 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2fql7" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.151849 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.152664 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.202210 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303154 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303209 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303235 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee71bf1-4264-4480-843b-04bef17b232a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303269 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303284 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303301 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d6af69e-ffd4-4001-b938-18df9e23981c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303339 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303375 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee71bf1-4264-4480-843b-04bef17b232a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303394 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303419 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303433 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303467 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303487 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303509 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303524 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303552 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303586 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlnl8\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-kube-api-access-xlnl8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303601 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303616 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303633 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnw85\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-kube-api-access-pnw85\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303652 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d6af69e-ffd4-4001-b938-18df9e23981c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.303674 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404605 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404864 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnw85\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-kube-api-access-pnw85\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404886 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d6af69e-ffd4-4001-b938-18df9e23981c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404914 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404946 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404963 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.404983 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee71bf1-4264-4480-843b-04bef17b232a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405004 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405020 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405036 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d6af69e-ffd4-4001-b938-18df9e23981c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405059 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405095 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee71bf1-4264-4480-843b-04bef17b232a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405116 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405138 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405153 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405178 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405205 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405227 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405242 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405273 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405310 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlnl8\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-kube-api-access-xlnl8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.405340 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.406152 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.408487 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.409296 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.409732 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d6af69e-ffd4-4001-b938-18df9e23981c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.410252 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.410892 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ee71bf1-4264-4480-843b-04bef17b232a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.410957 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.412127 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.412771 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ee71bf1-4264-4480-843b-04bef17b232a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.413116 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ee71bf1-4264-4480-843b-04bef17b232a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.413534 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.413561 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.413806 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.418020 4831 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.419583 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.420654 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.424067 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d6af69e-ffd4-4001-b938-18df9e23981c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.424618 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.427444 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d6af69e-ffd4-4001-b938-18df9e23981c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.430904 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.435889 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlnl8\" (UniqueName: \"kubernetes.io/projected/0ee71bf1-4264-4480-843b-04bef17b232a-kube-api-access-xlnl8\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.439268 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnw85\" (UniqueName: \"kubernetes.io/projected/3d6af69e-ffd4-4001-b938-18df9e23981c-kube-api-access-pnw85\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.447481 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ee71bf1-4264-4480-843b-04bef17b232a\") " pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.469632 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.471474 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"3d6af69e-ffd4-4001-b938-18df9e23981c\") " pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.738044 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.912286 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8b0fe50-d340-4379-8be9-4367a9c81d9f" path="/var/lib/kubelet/pods/e8b0fe50-d340-4379-8be9-4367a9c81d9f/volumes" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.913059 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0cc709e-e5c5-40fd-924f-aa3582689305" path="/var/lib/kubelet/pods/f0cc709e-e5c5-40fd-924f-aa3582689305/volumes" Nov 24 08:36:44 crc kubenswrapper[4831]: I1124 08:36:44.998797 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 24 08:36:45 crc kubenswrapper[4831]: W1124 08:36:45.230440 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d6af69e_ffd4_4001_b938_18df9e23981c.slice/crio-900cb42c75dd9807aa7bb2136caf5908c662b8fc04c72b3f204d83b9cc007af5 WatchSource:0}: Error finding container 900cb42c75dd9807aa7bb2136caf5908c662b8fc04c72b3f204d83b9cc007af5: Status 404 returned error can't find the container with id 900cb42c75dd9807aa7bb2136caf5908c662b8fc04c72b3f204d83b9cc007af5 Nov 24 08:36:45 crc kubenswrapper[4831]: I1124 08:36:45.233113 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 24 08:36:46 crc kubenswrapper[4831]: I1124 08:36:46.028931 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee71bf1-4264-4480-843b-04bef17b232a","Type":"ContainerStarted","Data":"e1158fa03e9bbfe2f49de537957391eed735ea7cb923c0df76bb22a070226a79"} Nov 24 08:36:46 crc kubenswrapper[4831]: I1124 08:36:46.030453 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d6af69e-ffd4-4001-b938-18df9e23981c","Type":"ContainerStarted","Data":"900cb42c75dd9807aa7bb2136caf5908c662b8fc04c72b3f204d83b9cc007af5"} Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.040842 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d6af69e-ffd4-4001-b938-18df9e23981c","Type":"ContainerStarted","Data":"4c7dfd3dc242beb80ad8fd43c5695862c0bd997f344db727088a1a9fe0fa7eb1"} Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.044886 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee71bf1-4264-4480-843b-04bef17b232a","Type":"ContainerStarted","Data":"409733ebab78373802a3c931261adf904ba4c45f0635828f4b5ced371f2083f8"} Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.159815 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.162072 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.168739 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.182591 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264487 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264565 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqw6d\" (UniqueName: \"kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264615 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264661 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264720 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.264760 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.366253 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.367184 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.367337 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.367404 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.367459 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368049 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368190 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368234 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368569 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368784 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.368865 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqw6d\" (UniqueName: \"kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.390255 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqw6d\" (UniqueName: \"kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d\") pod \"dnsmasq-dns-578b8d767c-kczj7\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:47 crc kubenswrapper[4831]: I1124 08:36:47.483542 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:48 crc kubenswrapper[4831]: I1124 08:36:48.180619 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:36:49 crc kubenswrapper[4831]: I1124 08:36:49.072058 4831 generic.go:334] "Generic (PLEG): container finished" podID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerID="a7cb62aedecb9948b25eea97dede88df104fc210c9cd3a0d2d90f5c7171ccbb2" exitCode=0 Nov 24 08:36:49 crc kubenswrapper[4831]: I1124 08:36:49.072118 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" event={"ID":"869f2ba7-6ce6-486b-9dd7-f2505f2b4515","Type":"ContainerDied","Data":"a7cb62aedecb9948b25eea97dede88df104fc210c9cd3a0d2d90f5c7171ccbb2"} Nov 24 08:36:49 crc kubenswrapper[4831]: I1124 08:36:49.072699 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" event={"ID":"869f2ba7-6ce6-486b-9dd7-f2505f2b4515","Type":"ContainerStarted","Data":"937f18c781490136aafb457632526b33bd64c4c1b33dc762dfc7d7712fc935cf"} Nov 24 08:36:50 crc kubenswrapper[4831]: I1124 08:36:50.087986 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" event={"ID":"869f2ba7-6ce6-486b-9dd7-f2505f2b4515","Type":"ContainerStarted","Data":"dacf93730e810e1ce2b263c80da3e12cb2f4b30683968a9df89e5ae6e1552041"} Nov 24 08:36:50 crc kubenswrapper[4831]: I1124 08:36:50.088680 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:50 crc kubenswrapper[4831]: I1124 08:36:50.113002 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" podStartSLOduration=3.112972317 podStartE2EDuration="3.112972317s" podCreationTimestamp="2025-11-24 08:36:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:36:50.104549389 +0000 UTC m=+1283.979694602" watchObservedRunningTime="2025-11-24 08:36:50.112972317 +0000 UTC m=+1283.988117470" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.485788 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.543305 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.543633 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="dnsmasq-dns" containerID="cri-o://2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772" gracePeriod=10 Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.712850 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-pw8bh"] Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.716434 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.750672 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-pw8bh"] Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.783825 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-config\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.783871 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-dns-svc\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.783909 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.784000 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.784022 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.784039 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkbr6\" (UniqueName: \"kubernetes.io/projected/af25e9d0-61ba-4758-8c40-1df95e214067-kube-api-access-vkbr6\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887172 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-config\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887223 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-dns-svc\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887271 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887367 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887393 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.887421 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkbr6\" (UniqueName: \"kubernetes.io/projected/af25e9d0-61ba-4758-8c40-1df95e214067-kube-api-access-vkbr6\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.888120 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-config\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.888776 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-nb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.890653 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-openstack-edpm-ipam\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.890857 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-ovsdbserver-sb\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.905480 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af25e9d0-61ba-4758-8c40-1df95e214067-dns-svc\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:57 crc kubenswrapper[4831]: I1124 08:36:57.916181 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkbr6\" (UniqueName: \"kubernetes.io/projected/af25e9d0-61ba-4758-8c40-1df95e214067-kube-api-access-vkbr6\") pod \"dnsmasq-dns-667ff9c869-pw8bh\" (UID: \"af25e9d0-61ba-4758-8c40-1df95e214067\") " pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.067144 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.171309 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.187063 4831 generic.go:334] "Generic (PLEG): container finished" podID="21835b63-320c-4149-b01d-d7afa343a341" containerID="2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772" exitCode=0 Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.187217 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.187272 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" event={"ID":"21835b63-320c-4149-b01d-d7afa343a341","Type":"ContainerDied","Data":"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772"} Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.187690 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-lrjz9" event={"ID":"21835b63-320c-4149-b01d-d7afa343a341","Type":"ContainerDied","Data":"81d1729848168c7b9f50f6b136e85212dfa2e1f13014b08a9f9f33510c47450a"} Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.187730 4831 scope.go:117] "RemoveContainer" containerID="2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.224546 4831 scope.go:117] "RemoveContainer" containerID="42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.271326 4831 scope.go:117] "RemoveContainer" containerID="2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772" Nov 24 08:36:58 crc kubenswrapper[4831]: E1124 08:36:58.273766 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772\": container with ID starting with 2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772 not found: ID does not exist" containerID="2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.273813 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772"} err="failed to get container status \"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772\": rpc error: code = NotFound desc = could not find container \"2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772\": container with ID starting with 2fbbdfa19a3b9567d51f0cadae6c513b65ea105c845fdbbabbd23a60c593e772 not found: ID does not exist" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.273843 4831 scope.go:117] "RemoveContainer" containerID="42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469" Nov 24 08:36:58 crc kubenswrapper[4831]: E1124 08:36:58.274136 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469\": container with ID starting with 42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469 not found: ID does not exist" containerID="42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.274187 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469"} err="failed to get container status \"42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469\": rpc error: code = NotFound desc = could not find container \"42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469\": container with ID starting with 42d7b2d04820d25fe82693bfb0ae20b47804dc8d70dc0ad2bdbe11c5a8ff7469 not found: ID does not exist" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.301528 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lll9p\" (UniqueName: \"kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p\") pod \"21835b63-320c-4149-b01d-d7afa343a341\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.301801 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config\") pod \"21835b63-320c-4149-b01d-d7afa343a341\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.301837 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc\") pod \"21835b63-320c-4149-b01d-d7afa343a341\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.301868 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb\") pod \"21835b63-320c-4149-b01d-d7afa343a341\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.301906 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb\") pod \"21835b63-320c-4149-b01d-d7afa343a341\" (UID: \"21835b63-320c-4149-b01d-d7afa343a341\") " Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.319078 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p" (OuterVolumeSpecName: "kube-api-access-lll9p") pod "21835b63-320c-4149-b01d-d7afa343a341" (UID: "21835b63-320c-4149-b01d-d7afa343a341"). InnerVolumeSpecName "kube-api-access-lll9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.386010 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config" (OuterVolumeSpecName: "config") pod "21835b63-320c-4149-b01d-d7afa343a341" (UID: "21835b63-320c-4149-b01d-d7afa343a341"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.395862 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21835b63-320c-4149-b01d-d7afa343a341" (UID: "21835b63-320c-4149-b01d-d7afa343a341"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.396179 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21835b63-320c-4149-b01d-d7afa343a341" (UID: "21835b63-320c-4149-b01d-d7afa343a341"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.397640 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21835b63-320c-4149-b01d-d7afa343a341" (UID: "21835b63-320c-4149-b01d-d7afa343a341"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.401794 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.401920 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.404775 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.404809 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.404824 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.404837 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lll9p\" (UniqueName: \"kubernetes.io/projected/21835b63-320c-4149-b01d-d7afa343a341-kube-api-access-lll9p\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.404848 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21835b63-320c-4149-b01d-d7afa343a341-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.519658 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.528270 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-lrjz9"] Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.592188 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-667ff9c869-pw8bh"] Nov 24 08:36:58 crc kubenswrapper[4831]: I1124 08:36:58.902719 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21835b63-320c-4149-b01d-d7afa343a341" path="/var/lib/kubelet/pods/21835b63-320c-4149-b01d-d7afa343a341/volumes" Nov 24 08:36:59 crc kubenswrapper[4831]: I1124 08:36:59.198214 4831 generic.go:334] "Generic (PLEG): container finished" podID="af25e9d0-61ba-4758-8c40-1df95e214067" containerID="62b8a99dcc983a26e6d55912b956bd44de657c3bc0c8369bf7e4633147826c2f" exitCode=0 Nov 24 08:36:59 crc kubenswrapper[4831]: I1124 08:36:59.198299 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" event={"ID":"af25e9d0-61ba-4758-8c40-1df95e214067","Type":"ContainerDied","Data":"62b8a99dcc983a26e6d55912b956bd44de657c3bc0c8369bf7e4633147826c2f"} Nov 24 08:36:59 crc kubenswrapper[4831]: I1124 08:36:59.198493 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" event={"ID":"af25e9d0-61ba-4758-8c40-1df95e214067","Type":"ContainerStarted","Data":"bc88a0d4f877e2732c5eb809289a5449f95aea9071c9676555c34eb78440ef45"} Nov 24 08:37:00 crc kubenswrapper[4831]: I1124 08:37:00.209675 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" event={"ID":"af25e9d0-61ba-4758-8c40-1df95e214067","Type":"ContainerStarted","Data":"d0cfcbae61828cf20ce39b8f7009c53790e7be596c3751e7fd434ccd91f7bd30"} Nov 24 08:37:00 crc kubenswrapper[4831]: I1124 08:37:00.211361 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:37:00 crc kubenswrapper[4831]: I1124 08:37:00.233605 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" podStartSLOduration=3.23358405 podStartE2EDuration="3.23358405s" podCreationTimestamp="2025-11-24 08:36:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:37:00.226033477 +0000 UTC m=+1294.101178650" watchObservedRunningTime="2025-11-24 08:37:00.23358405 +0000 UTC m=+1294.108729203" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.069084 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-667ff9c869-pw8bh" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.141453 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.141706 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="dnsmasq-dns" containerID="cri-o://dacf93730e810e1ce2b263c80da3e12cb2f4b30683968a9df89e5ae6e1552041" gracePeriod=10 Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.303305 4831 generic.go:334] "Generic (PLEG): container finished" podID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerID="dacf93730e810e1ce2b263c80da3e12cb2f4b30683968a9df89e5ae6e1552041" exitCode=0 Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.303374 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" event={"ID":"869f2ba7-6ce6-486b-9dd7-f2505f2b4515","Type":"ContainerDied","Data":"dacf93730e810e1ce2b263c80da3e12cb2f4b30683968a9df89e5ae6e1552041"} Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.644721 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.713934 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqw6d\" (UniqueName: \"kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.714041 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.714123 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.714159 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.714254 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.714299 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config\") pod \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\" (UID: \"869f2ba7-6ce6-486b-9dd7-f2505f2b4515\") " Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.747673 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d" (OuterVolumeSpecName: "kube-api-access-nqw6d") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "kube-api-access-nqw6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.771363 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config" (OuterVolumeSpecName: "config") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.794610 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.800865 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.809614 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.815915 4831 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.815942 4831 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.815952 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.815960 4831 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-config\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.815968 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqw6d\" (UniqueName: \"kubernetes.io/projected/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-kube-api-access-nqw6d\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.831351 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "869f2ba7-6ce6-486b-9dd7-f2505f2b4515" (UID: "869f2ba7-6ce6-486b-9dd7-f2505f2b4515"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:37:08 crc kubenswrapper[4831]: I1124 08:37:08.917935 4831 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/869f2ba7-6ce6-486b-9dd7-f2505f2b4515-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.317631 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" event={"ID":"869f2ba7-6ce6-486b-9dd7-f2505f2b4515","Type":"ContainerDied","Data":"937f18c781490136aafb457632526b33bd64c4c1b33dc762dfc7d7712fc935cf"} Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.317713 4831 scope.go:117] "RemoveContainer" containerID="dacf93730e810e1ce2b263c80da3e12cb2f4b30683968a9df89e5ae6e1552041" Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.319109 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-kczj7" Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.355039 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.356785 4831 scope.go:117] "RemoveContainer" containerID="a7cb62aedecb9948b25eea97dede88df104fc210c9cd3a0d2d90f5c7171ccbb2" Nov 24 08:37:09 crc kubenswrapper[4831]: I1124 08:37:09.367085 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-kczj7"] Nov 24 08:37:10 crc kubenswrapper[4831]: I1124 08:37:10.907256 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" path="/var/lib/kubelet/pods/869f2ba7-6ce6-486b-9dd7-f2505f2b4515/volumes" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.459856 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt"] Nov 24 08:37:18 crc kubenswrapper[4831]: E1124 08:37:18.475365 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="init" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475399 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="init" Nov 24 08:37:18 crc kubenswrapper[4831]: E1124 08:37:18.475420 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475442 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: E1124 08:37:18.475506 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="init" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475519 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="init" Nov 24 08:37:18 crc kubenswrapper[4831]: E1124 08:37:18.475561 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475574 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475866 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="21835b63-320c-4149-b01d-d7afa343a341" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.475892 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="869f2ba7-6ce6-486b-9dd7-f2505f2b4515" containerName="dnsmasq-dns" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.477249 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.481249 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.481847 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.481987 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.483852 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt"] Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.487297 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.504175 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.504582 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.504886 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcgd9\" (UniqueName: \"kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.504936 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.606028 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.606421 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.606673 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcgd9\" (UniqueName: \"kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.606861 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.625336 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.625386 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.625783 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.634894 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcgd9\" (UniqueName: \"kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:18 crc kubenswrapper[4831]: I1124 08:37:18.802925 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.359768 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt"] Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.439095 4831 generic.go:334] "Generic (PLEG): container finished" podID="3d6af69e-ffd4-4001-b938-18df9e23981c" containerID="4c7dfd3dc242beb80ad8fd43c5695862c0bd997f344db727088a1a9fe0fa7eb1" exitCode=0 Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.439199 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d6af69e-ffd4-4001-b938-18df9e23981c","Type":"ContainerDied","Data":"4c7dfd3dc242beb80ad8fd43c5695862c0bd997f344db727088a1a9fe0fa7eb1"} Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.444883 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" event={"ID":"f56f5009-34b0-45e5-b303-1c7b123e7600","Type":"ContainerStarted","Data":"58ecd8cd788e299c9ba72b512f11b39e6d9800040be7ecbc13c9a8384140dc4c"} Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.448479 4831 generic.go:334] "Generic (PLEG): container finished" podID="0ee71bf1-4264-4480-843b-04bef17b232a" containerID="409733ebab78373802a3c931261adf904ba4c45f0635828f4b5ced371f2083f8" exitCode=0 Nov 24 08:37:19 crc kubenswrapper[4831]: I1124 08:37:19.448548 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee71bf1-4264-4480-843b-04bef17b232a","Type":"ContainerDied","Data":"409733ebab78373802a3c931261adf904ba4c45f0635828f4b5ced371f2083f8"} Nov 24 08:37:20 crc kubenswrapper[4831]: I1124 08:37:20.465752 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d6af69e-ffd4-4001-b938-18df9e23981c","Type":"ContainerStarted","Data":"903f3d6fb746beaf1ddb583bf118b7d2f06c28369f9dc4e3682fca8e0ebc15e1"} Nov 24 08:37:20 crc kubenswrapper[4831]: I1124 08:37:20.469452 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ee71bf1-4264-4480-843b-04bef17b232a","Type":"ContainerStarted","Data":"c549183482ebf3f4136cbdf2adfca6c37a623b18024a97076ba1940cc1749030"} Nov 24 08:37:20 crc kubenswrapper[4831]: I1124 08:37:20.469806 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:37:20 crc kubenswrapper[4831]: I1124 08:37:20.496372 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.496344925 podStartE2EDuration="36.496344925s" podCreationTimestamp="2025-11-24 08:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:37:20.487522986 +0000 UTC m=+1314.362668149" watchObservedRunningTime="2025-11-24 08:37:20.496344925 +0000 UTC m=+1314.371490108" Nov 24 08:37:20 crc kubenswrapper[4831]: I1124 08:37:20.519993 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.519976513 podStartE2EDuration="36.519976513s" podCreationTimestamp="2025-11-24 08:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:37:20.517010409 +0000 UTC m=+1314.392155582" watchObservedRunningTime="2025-11-24 08:37:20.519976513 +0000 UTC m=+1314.395121666" Nov 24 08:37:24 crc kubenswrapper[4831]: I1124 08:37:24.739818 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.401645 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.401980 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.402033 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.402980 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.403055 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e" gracePeriod=600 Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.554933 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e" exitCode=0 Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.554987 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e"} Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.555032 4831 scope.go:117] "RemoveContainer" containerID="5a1158cbfe2bf03b364c0a166a6c5178e1bd392926adafe2539e66130dfb8899" Nov 24 08:37:28 crc kubenswrapper[4831]: I1124 08:37:28.966231 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:37:29 crc kubenswrapper[4831]: I1124 08:37:29.571657 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8"} Nov 24 08:37:29 crc kubenswrapper[4831]: I1124 08:37:29.595909 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" event={"ID":"f56f5009-34b0-45e5-b303-1c7b123e7600","Type":"ContainerStarted","Data":"bd816e230472cfc4a9863a4d10d30b55e9e0f8b7eb95514e24cab7b73b8378ed"} Nov 24 08:37:29 crc kubenswrapper[4831]: I1124 08:37:29.651840 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" podStartSLOduration=2.055644521 podStartE2EDuration="11.651809392s" podCreationTimestamp="2025-11-24 08:37:18 +0000 UTC" firstStartedPulling="2025-11-24 08:37:19.367645566 +0000 UTC m=+1313.242790719" lastFinishedPulling="2025-11-24 08:37:28.963810437 +0000 UTC m=+1322.838955590" observedRunningTime="2025-11-24 08:37:29.62486642 +0000 UTC m=+1323.500011573" watchObservedRunningTime="2025-11-24 08:37:29.651809392 +0000 UTC m=+1323.526954545" Nov 24 08:37:34 crc kubenswrapper[4831]: I1124 08:37:34.474539 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 24 08:37:34 crc kubenswrapper[4831]: I1124 08:37:34.745729 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 24 08:37:40 crc kubenswrapper[4831]: I1124 08:37:40.705075 4831 generic.go:334] "Generic (PLEG): container finished" podID="f56f5009-34b0-45e5-b303-1c7b123e7600" containerID="bd816e230472cfc4a9863a4d10d30b55e9e0f8b7eb95514e24cab7b73b8378ed" exitCode=0 Nov 24 08:37:40 crc kubenswrapper[4831]: I1124 08:37:40.705225 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" event={"ID":"f56f5009-34b0-45e5-b303-1c7b123e7600","Type":"ContainerDied","Data":"bd816e230472cfc4a9863a4d10d30b55e9e0f8b7eb95514e24cab7b73b8378ed"} Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.150569 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.263518 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory\") pod \"f56f5009-34b0-45e5-b303-1c7b123e7600\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.264049 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcgd9\" (UniqueName: \"kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9\") pod \"f56f5009-34b0-45e5-b303-1c7b123e7600\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.264110 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key\") pod \"f56f5009-34b0-45e5-b303-1c7b123e7600\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.264149 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle\") pod \"f56f5009-34b0-45e5-b303-1c7b123e7600\" (UID: \"f56f5009-34b0-45e5-b303-1c7b123e7600\") " Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.269581 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9" (OuterVolumeSpecName: "kube-api-access-mcgd9") pod "f56f5009-34b0-45e5-b303-1c7b123e7600" (UID: "f56f5009-34b0-45e5-b303-1c7b123e7600"). InnerVolumeSpecName "kube-api-access-mcgd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.270434 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f56f5009-34b0-45e5-b303-1c7b123e7600" (UID: "f56f5009-34b0-45e5-b303-1c7b123e7600"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.304194 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory" (OuterVolumeSpecName: "inventory") pod "f56f5009-34b0-45e5-b303-1c7b123e7600" (UID: "f56f5009-34b0-45e5-b303-1c7b123e7600"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.319621 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f56f5009-34b0-45e5-b303-1c7b123e7600" (UID: "f56f5009-34b0-45e5-b303-1c7b123e7600"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.366586 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.366624 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcgd9\" (UniqueName: \"kubernetes.io/projected/f56f5009-34b0-45e5-b303-1c7b123e7600-kube-api-access-mcgd9\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.366635 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.366647 4831 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f56f5009-34b0-45e5-b303-1c7b123e7600-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.726692 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" event={"ID":"f56f5009-34b0-45e5-b303-1c7b123e7600","Type":"ContainerDied","Data":"58ecd8cd788e299c9ba72b512f11b39e6d9800040be7ecbc13c9a8384140dc4c"} Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.726745 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58ecd8cd788e299c9ba72b512f11b39e6d9800040be7ecbc13c9a8384140dc4c" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.726775 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.918152 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl"] Nov 24 08:37:42 crc kubenswrapper[4831]: E1124 08:37:42.918649 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f5009-34b0-45e5-b303-1c7b123e7600" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.918676 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f5009-34b0-45e5-b303-1c7b123e7600" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.918898 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56f5009-34b0-45e5-b303-1c7b123e7600" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.919535 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.926224 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.926899 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.928088 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.928760 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.945839 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl"] Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.976992 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.977108 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdbpt\" (UniqueName: \"kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.977131 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:42 crc kubenswrapper[4831]: I1124 08:37:42.977188 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.079289 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdbpt\" (UniqueName: \"kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.079345 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.079391 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.079466 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.083378 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.085395 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.090554 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.103870 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdbpt\" (UniqueName: \"kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.243477 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:37:43 crc kubenswrapper[4831]: I1124 08:37:43.844530 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl"] Nov 24 08:37:44 crc kubenswrapper[4831]: I1124 08:37:44.747015 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" event={"ID":"fcc403cb-362a-408c-a86b-842ee0b5f9f7","Type":"ContainerStarted","Data":"e219849429833baddc0be40e729292c445fc70270eec06267d5f46d60b5beeee"} Nov 24 08:37:44 crc kubenswrapper[4831]: I1124 08:37:44.747512 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" event={"ID":"fcc403cb-362a-408c-a86b-842ee0b5f9f7","Type":"ContainerStarted","Data":"76bc1804afb63d21ecff81120072f1b5b20982ec46395b1246e464c81943bdb8"} Nov 24 08:37:44 crc kubenswrapper[4831]: I1124 08:37:44.766579 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" podStartSLOduration=2.329394876 podStartE2EDuration="2.766556081s" podCreationTimestamp="2025-11-24 08:37:42 +0000 UTC" firstStartedPulling="2025-11-24 08:37:43.854938887 +0000 UTC m=+1337.730084050" lastFinishedPulling="2025-11-24 08:37:44.292100102 +0000 UTC m=+1338.167245255" observedRunningTime="2025-11-24 08:37:44.76614501 +0000 UTC m=+1338.641290173" watchObservedRunningTime="2025-11-24 08:37:44.766556081 +0000 UTC m=+1338.641701284" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.772937 4831 scope.go:117] "RemoveContainer" containerID="351e96e33d34f1dd7a1d56c166ca6a8cefdb3058b5fba117be0e7a2f7ffb73d5" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.821353 4831 scope.go:117] "RemoveContainer" containerID="46eee5fa1cbc845a0637d7078d8ffb82868b46d45a3de566cb29ba954b534461" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.857583 4831 scope.go:117] "RemoveContainer" containerID="474b53112447ea9aedc0b761e2bae117da65c20019e0c99f838485074bc9d4d9" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.898151 4831 scope.go:117] "RemoveContainer" containerID="6a1f253dccd2eceb4b96368b8a5e1b2330dfc1c9a7d3318435c861f05c2d1f8c" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.924465 4831 scope.go:117] "RemoveContainer" containerID="cd2098e6b3f498379b7b998f5714d4e29a6764be559af2a2f95b22ad2487511a" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.964687 4831 scope.go:117] "RemoveContainer" containerID="3ba4a3ec89f5016354761a2640f7a0c6338795895ff716e40ec3cb693d12d062" Nov 24 08:38:01 crc kubenswrapper[4831]: I1124 08:38:01.983963 4831 scope.go:117] "RemoveContainer" containerID="1b8fa3bb649db213deb7ccefb70efb72e246da20ebdaf6fd1854e146a3b41f07" Nov 24 08:38:02 crc kubenswrapper[4831]: I1124 08:38:02.006187 4831 scope.go:117] "RemoveContainer" containerID="7c831df1604165a09ea1c5798d73a1ef09c377057c5c164dd214cbd9b440ed18" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.730561 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.733464 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.749695 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.897481 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.897611 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7k9\" (UniqueName: \"kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.897647 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.999412 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7k9\" (UniqueName: \"kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.999539 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:46 crc kubenswrapper[4831]: I1124 08:38:46.999614 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:47 crc kubenswrapper[4831]: I1124 08:38:47.001108 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:47 crc kubenswrapper[4831]: I1124 08:38:47.001372 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:47 crc kubenswrapper[4831]: I1124 08:38:47.020379 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7k9\" (UniqueName: \"kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9\") pod \"redhat-operators-rrvgh\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:47 crc kubenswrapper[4831]: I1124 08:38:47.055844 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:47 crc kubenswrapper[4831]: I1124 08:38:47.510254 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:38:48 crc kubenswrapper[4831]: I1124 08:38:48.419873 4831 generic.go:334] "Generic (PLEG): container finished" podID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerID="6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002" exitCode=0 Nov 24 08:38:48 crc kubenswrapper[4831]: I1124 08:38:48.420008 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerDied","Data":"6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002"} Nov 24 08:38:48 crc kubenswrapper[4831]: I1124 08:38:48.420180 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerStarted","Data":"9a9445377a2678a38f1158387cc00e9182ed76f6a3109a9ae2409b1484fc36fc"} Nov 24 08:38:50 crc kubenswrapper[4831]: I1124 08:38:50.442001 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerStarted","Data":"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7"} Nov 24 08:38:54 crc kubenswrapper[4831]: I1124 08:38:54.485524 4831 generic.go:334] "Generic (PLEG): container finished" podID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerID="ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7" exitCode=0 Nov 24 08:38:54 crc kubenswrapper[4831]: I1124 08:38:54.485615 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerDied","Data":"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7"} Nov 24 08:38:55 crc kubenswrapper[4831]: I1124 08:38:55.498828 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerStarted","Data":"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd"} Nov 24 08:38:55 crc kubenswrapper[4831]: I1124 08:38:55.531959 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rrvgh" podStartSLOduration=3.014596008 podStartE2EDuration="9.531940274s" podCreationTimestamp="2025-11-24 08:38:46 +0000 UTC" firstStartedPulling="2025-11-24 08:38:48.424163391 +0000 UTC m=+1402.299308544" lastFinishedPulling="2025-11-24 08:38:54.941507657 +0000 UTC m=+1408.816652810" observedRunningTime="2025-11-24 08:38:55.517152076 +0000 UTC m=+1409.392297249" watchObservedRunningTime="2025-11-24 08:38:55.531940274 +0000 UTC m=+1409.407085427" Nov 24 08:38:57 crc kubenswrapper[4831]: I1124 08:38:57.057394 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:57 crc kubenswrapper[4831]: I1124 08:38:57.057823 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:38:58 crc kubenswrapper[4831]: I1124 08:38:58.101764 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrvgh" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" probeResult="failure" output=< Nov 24 08:38:58 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:38:58 crc kubenswrapper[4831]: > Nov 24 08:39:02 crc kubenswrapper[4831]: I1124 08:39:02.149783 4831 scope.go:117] "RemoveContainer" containerID="0ce1190f12c5cf3014548a6949bc4eb34b5c6220a376f40766c7eaf59f2f90d0" Nov 24 08:39:02 crc kubenswrapper[4831]: I1124 08:39:02.194826 4831 scope.go:117] "RemoveContainer" containerID="60f3f66a9016ea5a56cf062f8cb3b4f53c6051dbf6b63cf2e933cbb34852e775" Nov 24 08:39:08 crc kubenswrapper[4831]: I1124 08:39:08.105872 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrvgh" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" probeResult="failure" output=< Nov 24 08:39:08 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:39:08 crc kubenswrapper[4831]: > Nov 24 08:39:17 crc kubenswrapper[4831]: I1124 08:39:17.104267 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:39:17 crc kubenswrapper[4831]: I1124 08:39:17.152915 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:39:17 crc kubenswrapper[4831]: I1124 08:39:17.934929 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:39:18 crc kubenswrapper[4831]: I1124 08:39:18.738087 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rrvgh" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" containerID="cri-o://8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd" gracePeriod=2 Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.145907 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.212735 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content\") pod \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.212847 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities\") pod \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.212921 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc7k9\" (UniqueName: \"kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9\") pod \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\" (UID: \"5e312ff0-27c4-416d-8b3e-5f9cfe68d416\") " Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.213477 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities" (OuterVolumeSpecName: "utilities") pod "5e312ff0-27c4-416d-8b3e-5f9cfe68d416" (UID: "5e312ff0-27c4-416d-8b3e-5f9cfe68d416"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.217785 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9" (OuterVolumeSpecName: "kube-api-access-pc7k9") pod "5e312ff0-27c4-416d-8b3e-5f9cfe68d416" (UID: "5e312ff0-27c4-416d-8b3e-5f9cfe68d416"). InnerVolumeSpecName "kube-api-access-pc7k9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.312559 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e312ff0-27c4-416d-8b3e-5f9cfe68d416" (UID: "5e312ff0-27c4-416d-8b3e-5f9cfe68d416"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.315312 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.315343 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.315356 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc7k9\" (UniqueName: \"kubernetes.io/projected/5e312ff0-27c4-416d-8b3e-5f9cfe68d416-kube-api-access-pc7k9\") on node \"crc\" DevicePath \"\"" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.749758 4831 generic.go:334] "Generic (PLEG): container finished" podID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerID="8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd" exitCode=0 Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.749812 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerDied","Data":"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd"} Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.749842 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrvgh" event={"ID":"5e312ff0-27c4-416d-8b3e-5f9cfe68d416","Type":"ContainerDied","Data":"9a9445377a2678a38f1158387cc00e9182ed76f6a3109a9ae2409b1484fc36fc"} Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.749887 4831 scope.go:117] "RemoveContainer" containerID="8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.750004 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrvgh" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.781830 4831 scope.go:117] "RemoveContainer" containerID="ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.809897 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.821985 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rrvgh"] Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.822769 4831 scope.go:117] "RemoveContainer" containerID="6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.871969 4831 scope.go:117] "RemoveContainer" containerID="8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd" Nov 24 08:39:19 crc kubenswrapper[4831]: E1124 08:39:19.872531 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd\": container with ID starting with 8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd not found: ID does not exist" containerID="8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.872580 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd"} err="failed to get container status \"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd\": rpc error: code = NotFound desc = could not find container \"8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd\": container with ID starting with 8449a3f42c7674426a421df598cf28133161ea4b6a890d234366a4df0fad03cd not found: ID does not exist" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.872617 4831 scope.go:117] "RemoveContainer" containerID="ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7" Nov 24 08:39:19 crc kubenswrapper[4831]: E1124 08:39:19.873112 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7\": container with ID starting with ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7 not found: ID does not exist" containerID="ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.873159 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7"} err="failed to get container status \"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7\": rpc error: code = NotFound desc = could not find container \"ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7\": container with ID starting with ac8957eefc0fc210dc5d6ce84c1490d0d80f6f99d67f0a8d7a9a6f9befd765c7 not found: ID does not exist" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.873190 4831 scope.go:117] "RemoveContainer" containerID="6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002" Nov 24 08:39:19 crc kubenswrapper[4831]: E1124 08:39:19.873556 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002\": container with ID starting with 6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002 not found: ID does not exist" containerID="6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002" Nov 24 08:39:19 crc kubenswrapper[4831]: I1124 08:39:19.873585 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002"} err="failed to get container status \"6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002\": rpc error: code = NotFound desc = could not find container \"6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002\": container with ID starting with 6894e87ed622c18fb7c605ec687a32cf0cdba89a30de0b1016c2380c52318002 not found: ID does not exist" Nov 24 08:39:20 crc kubenswrapper[4831]: I1124 08:39:20.904906 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" path="/var/lib/kubelet/pods/5e312ff0-27c4-416d-8b3e-5f9cfe68d416/volumes" Nov 24 08:39:58 crc kubenswrapper[4831]: I1124 08:39:58.401730 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:39:58 crc kubenswrapper[4831]: I1124 08:39:58.402677 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.043196 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:16 crc kubenswrapper[4831]: E1124 08:40:16.048111 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="extract-utilities" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.048258 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="extract-utilities" Nov 24 08:40:16 crc kubenswrapper[4831]: E1124 08:40:16.048355 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="extract-content" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.048447 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="extract-content" Nov 24 08:40:16 crc kubenswrapper[4831]: E1124 08:40:16.048542 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.048609 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.048925 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e312ff0-27c4-416d-8b3e-5f9cfe68d416" containerName="registry-server" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.050594 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.062789 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.214501 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvnj\" (UniqueName: \"kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.214976 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.215158 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.316393 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.316760 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvnj\" (UniqueName: \"kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.316931 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.317109 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.317608 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.336499 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvnj\" (UniqueName: \"kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj\") pod \"certified-operators-6j7pd\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.409506 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:16 crc kubenswrapper[4831]: I1124 08:40:16.857787 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:17 crc kubenswrapper[4831]: I1124 08:40:17.262075 4831 generic.go:334] "Generic (PLEG): container finished" podID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerID="83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67" exitCode=0 Nov 24 08:40:17 crc kubenswrapper[4831]: I1124 08:40:17.262209 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerDied","Data":"83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67"} Nov 24 08:40:17 crc kubenswrapper[4831]: I1124 08:40:17.262681 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerStarted","Data":"7c82e184dc159b7878865b50465585f25bb214d3aec8f77d2d40dee020997219"} Nov 24 08:40:17 crc kubenswrapper[4831]: I1124 08:40:17.266267 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:40:18 crc kubenswrapper[4831]: I1124 08:40:18.276089 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerStarted","Data":"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b"} Nov 24 08:40:20 crc kubenswrapper[4831]: I1124 08:40:20.297289 4831 generic.go:334] "Generic (PLEG): container finished" podID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerID="f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b" exitCode=0 Nov 24 08:40:20 crc kubenswrapper[4831]: I1124 08:40:20.297346 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerDied","Data":"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b"} Nov 24 08:40:21 crc kubenswrapper[4831]: I1124 08:40:21.309098 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerStarted","Data":"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb"} Nov 24 08:40:21 crc kubenswrapper[4831]: I1124 08:40:21.332725 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6j7pd" podStartSLOduration=1.727331291 podStartE2EDuration="5.332706332s" podCreationTimestamp="2025-11-24 08:40:16 +0000 UTC" firstStartedPulling="2025-11-24 08:40:17.265830397 +0000 UTC m=+1491.140975550" lastFinishedPulling="2025-11-24 08:40:20.871205438 +0000 UTC m=+1494.746350591" observedRunningTime="2025-11-24 08:40:21.329780008 +0000 UTC m=+1495.204925191" watchObservedRunningTime="2025-11-24 08:40:21.332706332 +0000 UTC m=+1495.207851485" Nov 24 08:40:26 crc kubenswrapper[4831]: I1124 08:40:26.410612 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:26 crc kubenswrapper[4831]: I1124 08:40:26.411081 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:26 crc kubenswrapper[4831]: I1124 08:40:26.459336 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:27 crc kubenswrapper[4831]: I1124 08:40:27.430412 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:27 crc kubenswrapper[4831]: I1124 08:40:27.496154 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:28 crc kubenswrapper[4831]: I1124 08:40:28.401766 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:40:28 crc kubenswrapper[4831]: I1124 08:40:28.401828 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:40:29 crc kubenswrapper[4831]: I1124 08:40:29.393338 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6j7pd" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="registry-server" containerID="cri-o://1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb" gracePeriod=2 Nov 24 08:40:29 crc kubenswrapper[4831]: I1124 08:40:29.887372 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.079663 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvvnj\" (UniqueName: \"kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj\") pod \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.079843 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content\") pod \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.079927 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities\") pod \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\" (UID: \"bb448bbc-738c-4e5a-a0b2-27883af3fef5\") " Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.080780 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities" (OuterVolumeSpecName: "utilities") pod "bb448bbc-738c-4e5a-a0b2-27883af3fef5" (UID: "bb448bbc-738c-4e5a-a0b2-27883af3fef5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.084941 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj" (OuterVolumeSpecName: "kube-api-access-vvvnj") pod "bb448bbc-738c-4e5a-a0b2-27883af3fef5" (UID: "bb448bbc-738c-4e5a-a0b2-27883af3fef5"). InnerVolumeSpecName "kube-api-access-vvvnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.143740 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb448bbc-738c-4e5a-a0b2-27883af3fef5" (UID: "bb448bbc-738c-4e5a-a0b2-27883af3fef5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.181716 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.182560 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb448bbc-738c-4e5a-a0b2-27883af3fef5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.182781 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvvnj\" (UniqueName: \"kubernetes.io/projected/bb448bbc-738c-4e5a-a0b2-27883af3fef5-kube-api-access-vvvnj\") on node \"crc\" DevicePath \"\"" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.403411 4831 generic.go:334] "Generic (PLEG): container finished" podID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerID="1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb" exitCode=0 Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.403467 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerDied","Data":"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb"} Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.403501 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6j7pd" event={"ID":"bb448bbc-738c-4e5a-a0b2-27883af3fef5","Type":"ContainerDied","Data":"7c82e184dc159b7878865b50465585f25bb214d3aec8f77d2d40dee020997219"} Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.403520 4831 scope.go:117] "RemoveContainer" containerID="1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.403708 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6j7pd" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.437440 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.438994 4831 scope.go:117] "RemoveContainer" containerID="f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.464945 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6j7pd"] Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.473382 4831 scope.go:117] "RemoveContainer" containerID="83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.521732 4831 scope.go:117] "RemoveContainer" containerID="1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb" Nov 24 08:40:30 crc kubenswrapper[4831]: E1124 08:40:30.522291 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb\": container with ID starting with 1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb not found: ID does not exist" containerID="1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.522376 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb"} err="failed to get container status \"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb\": rpc error: code = NotFound desc = could not find container \"1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb\": container with ID starting with 1cd0e3515e5a3851e8e3544bdf54c3a80a0b46a0c51075b5593973112a56cdbb not found: ID does not exist" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.522398 4831 scope.go:117] "RemoveContainer" containerID="f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b" Nov 24 08:40:30 crc kubenswrapper[4831]: E1124 08:40:30.522604 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b\": container with ID starting with f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b not found: ID does not exist" containerID="f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.522626 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b"} err="failed to get container status \"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b\": rpc error: code = NotFound desc = could not find container \"f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b\": container with ID starting with f598711bd0fff8e93284b58333cf11598e42394ed80766c1bc9a884568f9277b not found: ID does not exist" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.522638 4831 scope.go:117] "RemoveContainer" containerID="83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67" Nov 24 08:40:30 crc kubenswrapper[4831]: E1124 08:40:30.522880 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67\": container with ID starting with 83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67 not found: ID does not exist" containerID="83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.522899 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67"} err="failed to get container status \"83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67\": rpc error: code = NotFound desc = could not find container \"83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67\": container with ID starting with 83ac1c96d98d631142edbfcf9fbc9046994ee5d614786f97eddb0f53b1a3ec67 not found: ID does not exist" Nov 24 08:40:30 crc kubenswrapper[4831]: I1124 08:40:30.908213 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" path="/var/lib/kubelet/pods/bb448bbc-738c-4e5a-a0b2-27883af3fef5/volumes" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.401548 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.402076 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.402119 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.402827 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.402892 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" gracePeriod=600 Nov 24 08:40:58 crc kubenswrapper[4831]: E1124 08:40:58.526392 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.657878 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" exitCode=0 Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.657983 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8"} Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.659271 4831 scope.go:117] "RemoveContainer" containerID="d0d65ec3a66dc9f9ac8b45fb4d894c73d4fedb303fd5a35fa06e9e932c72ed7e" Nov 24 08:40:58 crc kubenswrapper[4831]: I1124 08:40:58.659826 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:40:58 crc kubenswrapper[4831]: E1124 08:40:58.660551 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.375834 4831 scope.go:117] "RemoveContainer" containerID="c64e6fd12cf514241c834e3d9f449c2e4758b32856e47c5d2b4db7566f5c0cdd" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.399175 4831 scope.go:117] "RemoveContainer" containerID="058b4381d164c1e31d21195c601da866731483fa997f882ddd74acce5522639b" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.419458 4831 scope.go:117] "RemoveContainer" containerID="e39f033bcc7de8d23a03832a19c362479924e6568ace2b408c2151f5a8a816ac" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.452972 4831 scope.go:117] "RemoveContainer" containerID="4a5ed1fb1f17b795132bc60e9852405f12f91249a8ecb2e6776d61061ee5b25b" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.479534 4831 scope.go:117] "RemoveContainer" containerID="4883f184578c131b7aea2b549436953ef6a4a1eb55f1d9a1053ca88e7faf8bfd" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.500678 4831 scope.go:117] "RemoveContainer" containerID="38ac015fbf7f9512a409b657f175e1e635b57d0593706072dd06e01f47e139b9" Nov 24 08:41:02 crc kubenswrapper[4831]: I1124 08:41:02.532008 4831 scope.go:117] "RemoveContainer" containerID="6ae768aa3b739a7e4daa33dc68d529f0a22bd59d873d4021bd75921fffd7ba5b" Nov 24 08:41:09 crc kubenswrapper[4831]: I1124 08:41:09.893499 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:41:09 crc kubenswrapper[4831]: E1124 08:41:09.894372 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:41:14 crc kubenswrapper[4831]: I1124 08:41:14.806502 4831 generic.go:334] "Generic (PLEG): container finished" podID="fcc403cb-362a-408c-a86b-842ee0b5f9f7" containerID="e219849429833baddc0be40e729292c445fc70270eec06267d5f46d60b5beeee" exitCode=0 Nov 24 08:41:14 crc kubenswrapper[4831]: I1124 08:41:14.806582 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" event={"ID":"fcc403cb-362a-408c-a86b-842ee0b5f9f7","Type":"ContainerDied","Data":"e219849429833baddc0be40e729292c445fc70270eec06267d5f46d60b5beeee"} Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.213905 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.285544 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle\") pod \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.285607 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdbpt\" (UniqueName: \"kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt\") pod \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.285688 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key\") pod \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.285759 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory\") pod \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\" (UID: \"fcc403cb-362a-408c-a86b-842ee0b5f9f7\") " Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.293016 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt" (OuterVolumeSpecName: "kube-api-access-cdbpt") pod "fcc403cb-362a-408c-a86b-842ee0b5f9f7" (UID: "fcc403cb-362a-408c-a86b-842ee0b5f9f7"). InnerVolumeSpecName "kube-api-access-cdbpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.311524 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "fcc403cb-362a-408c-a86b-842ee0b5f9f7" (UID: "fcc403cb-362a-408c-a86b-842ee0b5f9f7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.317732 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fcc403cb-362a-408c-a86b-842ee0b5f9f7" (UID: "fcc403cb-362a-408c-a86b-842ee0b5f9f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.325723 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory" (OuterVolumeSpecName: "inventory") pod "fcc403cb-362a-408c-a86b-842ee0b5f9f7" (UID: "fcc403cb-362a-408c-a86b-842ee0b5f9f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.387475 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.387507 4831 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.387517 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdbpt\" (UniqueName: \"kubernetes.io/projected/fcc403cb-362a-408c-a86b-842ee0b5f9f7-kube-api-access-cdbpt\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.387525 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fcc403cb-362a-408c-a86b-842ee0b5f9f7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.829184 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" event={"ID":"fcc403cb-362a-408c-a86b-842ee0b5f9f7","Type":"ContainerDied","Data":"76bc1804afb63d21ecff81120072f1b5b20982ec46395b1246e464c81943bdb8"} Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.829232 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76bc1804afb63d21ecff81120072f1b5b20982ec46395b1246e464c81943bdb8" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.829270 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.938131 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v"] Nov 24 08:41:16 crc kubenswrapper[4831]: E1124 08:41:16.938964 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc403cb-362a-408c-a86b-842ee0b5f9f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.939090 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc403cb-362a-408c-a86b-842ee0b5f9f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 08:41:16 crc kubenswrapper[4831]: E1124 08:41:16.939200 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="registry-server" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.939273 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="registry-server" Nov 24 08:41:16 crc kubenswrapper[4831]: E1124 08:41:16.939376 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="extract-content" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.939454 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="extract-content" Nov 24 08:41:16 crc kubenswrapper[4831]: E1124 08:41:16.939572 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="extract-utilities" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.939654 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="extract-utilities" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.939955 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc403cb-362a-408c-a86b-842ee0b5f9f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.940057 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb448bbc-738c-4e5a-a0b2-27883af3fef5" containerName="registry-server" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.943760 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:16 crc kubenswrapper[4831]: E1124 08:41:16.960099 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcc403cb_362a_408c_a86b_842ee0b5f9f7.slice/crio-76bc1804afb63d21ecff81120072f1b5b20982ec46395b1246e464c81943bdb8\": RecentStats: unable to find data in memory cache]" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.992887 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.995487 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v"] Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.997552 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.998668 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:41:16 crc kubenswrapper[4831]: I1124 08:41:16.999465 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.100898 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.101693 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vscng\" (UniqueName: \"kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.101730 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.203810 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vscng\" (UniqueName: \"kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.203888 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.204006 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.211920 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.222166 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.227552 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vscng\" (UniqueName: \"kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.301932 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.825485 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v"] Nov 24 08:41:17 crc kubenswrapper[4831]: I1124 08:41:17.838646 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" event={"ID":"867472cb-8f0a-4f19-88c5-090bb1165a22","Type":"ContainerStarted","Data":"cbc2354989167400f9037a23aca4c0c1a9213b6a4c5da5e2c07b74e320213e77"} Nov 24 08:41:18 crc kubenswrapper[4831]: I1124 08:41:18.852751 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" event={"ID":"867472cb-8f0a-4f19-88c5-090bb1165a22","Type":"ContainerStarted","Data":"034a4883440a9dc86378730d8669096319c18b2de1f994cef02f1e7b36adf5c3"} Nov 24 08:41:18 crc kubenswrapper[4831]: I1124 08:41:18.875113 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" podStartSLOduration=2.401803187 podStartE2EDuration="2.875088798s" podCreationTimestamp="2025-11-24 08:41:16 +0000 UTC" firstStartedPulling="2025-11-24 08:41:17.815248275 +0000 UTC m=+1551.690393428" lastFinishedPulling="2025-11-24 08:41:18.288533886 +0000 UTC m=+1552.163679039" observedRunningTime="2025-11-24 08:41:18.871840695 +0000 UTC m=+1552.746985868" watchObservedRunningTime="2025-11-24 08:41:18.875088798 +0000 UTC m=+1552.750233941" Nov 24 08:41:23 crc kubenswrapper[4831]: I1124 08:41:23.893813 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:41:23 crc kubenswrapper[4831]: E1124 08:41:23.895823 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:41:38 crc kubenswrapper[4831]: I1124 08:41:38.893971 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:41:38 crc kubenswrapper[4831]: E1124 08:41:38.894697 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.043152 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-rh8vp"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.052915 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-c8rqp"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.063266 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7923-account-create-tdcf5"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.069843 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-rh8vp"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.078659 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-99fa-account-create-vvht6"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.087097 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7923-account-create-tdcf5"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.094538 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-c8rqp"] Nov 24 08:41:39 crc kubenswrapper[4831]: I1124 08:41:39.101227 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-99fa-account-create-vvht6"] Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.030806 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-eb35-account-create-958fz"] Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.043416 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-eb35-account-create-958fz"] Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.053653 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-qdhq6"] Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.063915 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-qdhq6"] Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.904293 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2312deb6-6491-4cd3-8e15-967215e9d9e0" path="/var/lib/kubelet/pods/2312deb6-6491-4cd3-8e15-967215e9d9e0/volumes" Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.905218 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d7f359-ef83-48f5-8d48-8256acbd0045" path="/var/lib/kubelet/pods/35d7f359-ef83-48f5-8d48-8256acbd0045/volumes" Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.906031 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496f9648-9656-4eff-9589-03132e32cec7" path="/var/lib/kubelet/pods/496f9648-9656-4eff-9589-03132e32cec7/volumes" Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.906752 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d3545d9-9a47-49a5-8353-751829dfc507" path="/var/lib/kubelet/pods/6d3545d9-9a47-49a5-8353-751829dfc507/volumes" Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.914175 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc257041-e2dc-415c-8988-53de7db6a961" path="/var/lib/kubelet/pods/dc257041-e2dc-415c-8988-53de7db6a961/volumes" Nov 24 08:41:40 crc kubenswrapper[4831]: I1124 08:41:40.914970 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ece974ef-9247-4703-8277-65eab2031685" path="/var/lib/kubelet/pods/ece974ef-9247-4703-8277-65eab2031685/volumes" Nov 24 08:41:52 crc kubenswrapper[4831]: I1124 08:41:52.893903 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:41:52 crc kubenswrapper[4831]: E1124 08:41:52.894690 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.613838 4831 scope.go:117] "RemoveContainer" containerID="f3d9aeeacec416686d8f484558c2f7cda865a83318ff596b6ecb71abcc20cdba" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.643426 4831 scope.go:117] "RemoveContainer" containerID="c54af51b868f996359c36821fc7a715427cda6412e4a41061c91f6143d5ffc58" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.685146 4831 scope.go:117] "RemoveContainer" containerID="4efc22cca0bede96ca0107647c613a9ce44818b36ecb45bd03b404b1d678ff7b" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.725657 4831 scope.go:117] "RemoveContainer" containerID="44977e20af8ef71481775a5bb73eede1e929fadc5ece316252ed28125e636ecc" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.787484 4831 scope.go:117] "RemoveContainer" containerID="f4162bfa538d2eade51040e40a35de2e44d974b064e72d756044e40738b20e9e" Nov 24 08:42:02 crc kubenswrapper[4831]: I1124 08:42:02.810120 4831 scope.go:117] "RemoveContainer" containerID="28b6539b0a0e0ffb0c03bd8f5b3dd6ad1b5d70635a50d3b364f817c67ed5092d" Nov 24 08:42:03 crc kubenswrapper[4831]: I1124 08:42:03.893785 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:42:03 crc kubenswrapper[4831]: E1124 08:42:03.893994 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:42:10 crc kubenswrapper[4831]: I1124 08:42:10.037491 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qzcrh"] Nov 24 08:42:10 crc kubenswrapper[4831]: I1124 08:42:10.044013 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qzcrh"] Nov 24 08:42:10 crc kubenswrapper[4831]: I1124 08:42:10.920311 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9b9615-df1a-4680-8460-c150476edd94" path="/var/lib/kubelet/pods/0c9b9615-df1a-4680-8460-c150476edd94/volumes" Nov 24 08:42:17 crc kubenswrapper[4831]: I1124 08:42:17.044003 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-041b-account-create-xz99f"] Nov 24 08:42:17 crc kubenswrapper[4831]: I1124 08:42:17.055124 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-041b-account-create-xz99f"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.036356 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wk2jg"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.047059 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a381-account-create-5jxnk"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.054946 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wk2jg"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.066537 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xmkr7"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.077738 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a381-account-create-5jxnk"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.085172 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-daa7-account-create-rbtz7"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.094059 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-frfnc"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.103791 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xmkr7"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.110698 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-daa7-account-create-rbtz7"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.117659 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-frfnc"] Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.899398 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:42:18 crc kubenswrapper[4831]: E1124 08:42:18.899759 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.912734 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01fbdeba-6c1f-423e-92b0-6e69136756f1" path="/var/lib/kubelet/pods/01fbdeba-6c1f-423e-92b0-6e69136756f1/volumes" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.913761 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f3b2d59-b9aa-44ae-a7ca-2f597a08435e" path="/var/lib/kubelet/pods/2f3b2d59-b9aa-44ae-a7ca-2f597a08435e/volumes" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.914459 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e80d4e0-3649-4a08-b1ec-d606b6be82eb" path="/var/lib/kubelet/pods/3e80d4e0-3649-4a08-b1ec-d606b6be82eb/volumes" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.915134 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89bc8072-5ead-41da-9dec-790ab8cd3d58" path="/var/lib/kubelet/pods/89bc8072-5ead-41da-9dec-790ab8cd3d58/volumes" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.917225 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c01f8f81-1e14-40a3-a1ed-195d691d9fba" path="/var/lib/kubelet/pods/c01f8f81-1e14-40a3-a1ed-195d691d9fba/volumes" Nov 24 08:42:18 crc kubenswrapper[4831]: I1124 08:42:18.917907 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd0e715a-72c6-4240-9dac-fd3efb994bff" path="/var/lib/kubelet/pods/dd0e715a-72c6-4240-9dac-fd3efb994bff/volumes" Nov 24 08:42:27 crc kubenswrapper[4831]: I1124 08:42:27.036807 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-m95zj"] Nov 24 08:42:27 crc kubenswrapper[4831]: I1124 08:42:27.045583 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-m95zj"] Nov 24 08:42:28 crc kubenswrapper[4831]: I1124 08:42:28.905580 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="843cd907-38e7-401e-ac60-69a7c3b17610" path="/var/lib/kubelet/pods/843cd907-38e7-401e-ac60-69a7c3b17610/volumes" Nov 24 08:42:32 crc kubenswrapper[4831]: I1124 08:42:32.893562 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:42:32 crc kubenswrapper[4831]: E1124 08:42:32.895232 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:42:34 crc kubenswrapper[4831]: I1124 08:42:34.549136 4831 generic.go:334] "Generic (PLEG): container finished" podID="867472cb-8f0a-4f19-88c5-090bb1165a22" containerID="034a4883440a9dc86378730d8669096319c18b2de1f994cef02f1e7b36adf5c3" exitCode=0 Nov 24 08:42:34 crc kubenswrapper[4831]: I1124 08:42:34.549223 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" event={"ID":"867472cb-8f0a-4f19-88c5-090bb1165a22","Type":"ContainerDied","Data":"034a4883440a9dc86378730d8669096319c18b2de1f994cef02f1e7b36adf5c3"} Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.013640 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.146841 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory\") pod \"867472cb-8f0a-4f19-88c5-090bb1165a22\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.146990 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vscng\" (UniqueName: \"kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng\") pod \"867472cb-8f0a-4f19-88c5-090bb1165a22\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.147077 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key\") pod \"867472cb-8f0a-4f19-88c5-090bb1165a22\" (UID: \"867472cb-8f0a-4f19-88c5-090bb1165a22\") " Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.157552 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng" (OuterVolumeSpecName: "kube-api-access-vscng") pod "867472cb-8f0a-4f19-88c5-090bb1165a22" (UID: "867472cb-8f0a-4f19-88c5-090bb1165a22"). InnerVolumeSpecName "kube-api-access-vscng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.171737 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory" (OuterVolumeSpecName: "inventory") pod "867472cb-8f0a-4f19-88c5-090bb1165a22" (UID: "867472cb-8f0a-4f19-88c5-090bb1165a22"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.176685 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "867472cb-8f0a-4f19-88c5-090bb1165a22" (UID: "867472cb-8f0a-4f19-88c5-090bb1165a22"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.249729 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.249772 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/867472cb-8f0a-4f19-88c5-090bb1165a22-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.249782 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vscng\" (UniqueName: \"kubernetes.io/projected/867472cb-8f0a-4f19-88c5-090bb1165a22-kube-api-access-vscng\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.566127 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" event={"ID":"867472cb-8f0a-4f19-88c5-090bb1165a22","Type":"ContainerDied","Data":"cbc2354989167400f9037a23aca4c0c1a9213b6a4c5da5e2c07b74e320213e77"} Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.566174 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbc2354989167400f9037a23aca4c0c1a9213b6a4c5da5e2c07b74e320213e77" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.566196 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.648687 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq"] Nov 24 08:42:36 crc kubenswrapper[4831]: E1124 08:42:36.649147 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867472cb-8f0a-4f19-88c5-090bb1165a22" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.649170 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="867472cb-8f0a-4f19-88c5-090bb1165a22" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.649389 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="867472cb-8f0a-4f19-88c5-090bb1165a22" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.655642 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.659173 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.659212 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.659434 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.663010 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq"] Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.664963 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.758803 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429mf\" (UniqueName: \"kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.758862 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.758910 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.861124 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.861263 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429mf\" (UniqueName: \"kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.861297 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.863988 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.863988 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.881010 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429mf\" (UniqueName: \"kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:36 crc kubenswrapper[4831]: I1124 08:42:36.978820 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:37 crc kubenswrapper[4831]: I1124 08:42:37.513969 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq"] Nov 24 08:42:37 crc kubenswrapper[4831]: I1124 08:42:37.577777 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" event={"ID":"450b41f3-ecf5-4f17-a774-ade769d2fa1f","Type":"ContainerStarted","Data":"1a9704567f526165a578a4d30dadeeb7bd466cb7ac6ff86c6219958949e89d32"} Nov 24 08:42:38 crc kubenswrapper[4831]: I1124 08:42:38.588752 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" event={"ID":"450b41f3-ecf5-4f17-a774-ade769d2fa1f","Type":"ContainerStarted","Data":"dc3ad52458845fcd9f3c0ad35956f344881fa89795a199644fcdbc6ce691b019"} Nov 24 08:42:43 crc kubenswrapper[4831]: I1124 08:42:43.654531 4831 generic.go:334] "Generic (PLEG): container finished" podID="450b41f3-ecf5-4f17-a774-ade769d2fa1f" containerID="dc3ad52458845fcd9f3c0ad35956f344881fa89795a199644fcdbc6ce691b019" exitCode=0 Nov 24 08:42:43 crc kubenswrapper[4831]: I1124 08:42:43.654975 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" event={"ID":"450b41f3-ecf5-4f17-a774-ade769d2fa1f","Type":"ContainerDied","Data":"dc3ad52458845fcd9f3c0ad35956f344881fa89795a199644fcdbc6ce691b019"} Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.051538 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.227708 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory\") pod \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.227980 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429mf\" (UniqueName: \"kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf\") pod \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.228036 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key\") pod \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\" (UID: \"450b41f3-ecf5-4f17-a774-ade769d2fa1f\") " Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.233536 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf" (OuterVolumeSpecName: "kube-api-access-429mf") pod "450b41f3-ecf5-4f17-a774-ade769d2fa1f" (UID: "450b41f3-ecf5-4f17-a774-ade769d2fa1f"). InnerVolumeSpecName "kube-api-access-429mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.256185 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory" (OuterVolumeSpecName: "inventory") pod "450b41f3-ecf5-4f17-a774-ade769d2fa1f" (UID: "450b41f3-ecf5-4f17-a774-ade769d2fa1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.256925 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "450b41f3-ecf5-4f17-a774-ade769d2fa1f" (UID: "450b41f3-ecf5-4f17-a774-ade769d2fa1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.330937 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.330978 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429mf\" (UniqueName: \"kubernetes.io/projected/450b41f3-ecf5-4f17-a774-ade769d2fa1f-kube-api-access-429mf\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.330991 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/450b41f3-ecf5-4f17-a774-ade769d2fa1f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.680802 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" event={"ID":"450b41f3-ecf5-4f17-a774-ade769d2fa1f","Type":"ContainerDied","Data":"1a9704567f526165a578a4d30dadeeb7bd466cb7ac6ff86c6219958949e89d32"} Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.681113 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a9704567f526165a578a4d30dadeeb7bd466cb7ac6ff86c6219958949e89d32" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.680875 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.764580 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn"] Nov 24 08:42:45 crc kubenswrapper[4831]: E1124 08:42:45.765018 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="450b41f3-ecf5-4f17-a774-ade769d2fa1f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.765040 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="450b41f3-ecf5-4f17-a774-ade769d2fa1f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.765335 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="450b41f3-ecf5-4f17-a774-ade769d2fa1f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.766095 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.767966 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.768234 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.768570 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.773876 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.788780 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn"] Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.943762 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.943856 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vtqg\" (UniqueName: \"kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:45 crc kubenswrapper[4831]: I1124 08:42:45.944022 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.045484 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.045604 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vtqg\" (UniqueName: \"kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.045800 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.051964 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.052970 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.073691 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vtqg\" (UniqueName: \"kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jl4hn\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.095686 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.658910 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn"] Nov 24 08:42:46 crc kubenswrapper[4831]: I1124 08:42:46.692747 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" event={"ID":"23122f3c-2de2-4a17-b899-d71ea63b8b7f","Type":"ContainerStarted","Data":"c5ee9137f76b460522afb9fac979e6202095ba0d25fb3b353c037ddf338c83b3"} Nov 24 08:42:47 crc kubenswrapper[4831]: I1124 08:42:47.703287 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" event={"ID":"23122f3c-2de2-4a17-b899-d71ea63b8b7f","Type":"ContainerStarted","Data":"4ff51baa7b6f3f06eb373bf266f49c82bed20490f38a3850b038b88fc221de16"} Nov 24 08:42:47 crc kubenswrapper[4831]: I1124 08:42:47.728061 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" podStartSLOduration=2.29364111 podStartE2EDuration="2.728041767s" podCreationTimestamp="2025-11-24 08:42:45 +0000 UTC" firstStartedPulling="2025-11-24 08:42:46.654160353 +0000 UTC m=+1640.529305506" lastFinishedPulling="2025-11-24 08:42:47.088561 +0000 UTC m=+1640.963706163" observedRunningTime="2025-11-24 08:42:47.725027562 +0000 UTC m=+1641.600172725" watchObservedRunningTime="2025-11-24 08:42:47.728041767 +0000 UTC m=+1641.603186940" Nov 24 08:42:47 crc kubenswrapper[4831]: I1124 08:42:47.893516 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:42:47 crc kubenswrapper[4831]: E1124 08:42:47.893765 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:42:53 crc kubenswrapper[4831]: I1124 08:42:53.044747 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vm69f"] Nov 24 08:42:53 crc kubenswrapper[4831]: I1124 08:42:53.051801 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vm69f"] Nov 24 08:42:54 crc kubenswrapper[4831]: I1124 08:42:54.903037 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6303dec0-5e57-4012-b093-1307fd1211e9" path="/var/lib/kubelet/pods/6303dec0-5e57-4012-b093-1307fd1211e9/volumes" Nov 24 08:43:01 crc kubenswrapper[4831]: I1124 08:43:01.044720 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tv9wr"] Nov 24 08:43:01 crc kubenswrapper[4831]: I1124 08:43:01.065720 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tv9wr"] Nov 24 08:43:01 crc kubenswrapper[4831]: I1124 08:43:01.895380 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:43:01 crc kubenswrapper[4831]: E1124 08:43:01.895810 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.042475 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qqjkx"] Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.049867 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nzfjn"] Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.058248 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qqjkx"] Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.065542 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nzfjn"] Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.907365 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b702363c-0c9b-4804-a478-01ea54210d2c" path="/var/lib/kubelet/pods/b702363c-0c9b-4804-a478-01ea54210d2c/volumes" Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.908634 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b76e6599-c355-4d06-80bd-cb66824a64a4" path="/var/lib/kubelet/pods/b76e6599-c355-4d06-80bd-cb66824a64a4/volumes" Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.909829 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d698dc95-8dfe-483d-82ac-21584ce29659" path="/var/lib/kubelet/pods/d698dc95-8dfe-483d-82ac-21584ce29659/volumes" Nov 24 08:43:02 crc kubenswrapper[4831]: I1124 08:43:02.989935 4831 scope.go:117] "RemoveContainer" containerID="97fb9d8197a30d388f67688d72bd6be91d48f09824d8d96b2f68a14a31c2a3e8" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.023715 4831 scope.go:117] "RemoveContainer" containerID="6c3729170ea6d4c10fe9a03adf4076424673d06e6c77c517119f65470b208a2c" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.102963 4831 scope.go:117] "RemoveContainer" containerID="622430eb11da2ae1d25bcc31862089471717187021823e69cbfce98b462c515f" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.142380 4831 scope.go:117] "RemoveContainer" containerID="6058f3c92d28aabd97db5819d0be4d04e680ffcd16cf93e85d461a3d10c1f653" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.176546 4831 scope.go:117] "RemoveContainer" containerID="024e47611b52191ac038ef064f4a8addb1e41eafdcea32506c90c2ddbab87816" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.223851 4831 scope.go:117] "RemoveContainer" containerID="b38718cb79dbb610d00c953a9bdfd486986e181488e61a7e4e25567f74d50f36" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.283635 4831 scope.go:117] "RemoveContainer" containerID="4f5617adfd564206c13fa008d2a5a73da35553619760045bd1b6c9fe926cc4a8" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.313187 4831 scope.go:117] "RemoveContainer" containerID="8e1699bd4161a3cf314a2d2b35a8698dfc0a30ba026922b7233279171af08b08" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.333309 4831 scope.go:117] "RemoveContainer" containerID="03e5863499760678fbd083899acc363f4824eff8f205f899dec1e2724e5ab6db" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.355064 4831 scope.go:117] "RemoveContainer" containerID="34862a0c923d434d7c056566cd18b3c11d7da2a82dd454d85ca2aa6a7d1f696d" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.377308 4831 scope.go:117] "RemoveContainer" containerID="0236ca4597752c7d08649added42ad7ba113929de0b2b97eec87202ca72f1f0f" Nov 24 08:43:03 crc kubenswrapper[4831]: I1124 08:43:03.407432 4831 scope.go:117] "RemoveContainer" containerID="0056e5daaedbf3c446a0301f7940251ec01fa21fbba3f93937b2300d5e991173" Nov 24 08:43:13 crc kubenswrapper[4831]: I1124 08:43:13.893818 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:43:13 crc kubenswrapper[4831]: E1124 08:43:13.894684 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:43:23 crc kubenswrapper[4831]: I1124 08:43:23.039665 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tqhj9"] Nov 24 08:43:23 crc kubenswrapper[4831]: I1124 08:43:23.046366 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tqhj9"] Nov 24 08:43:24 crc kubenswrapper[4831]: I1124 08:43:24.907432 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a54561-e390-48c6-a27d-e516a013ebb3" path="/var/lib/kubelet/pods/46a54561-e390-48c6-a27d-e516a013ebb3/volumes" Nov 24 08:43:28 crc kubenswrapper[4831]: I1124 08:43:28.694868 4831 generic.go:334] "Generic (PLEG): container finished" podID="23122f3c-2de2-4a17-b899-d71ea63b8b7f" containerID="4ff51baa7b6f3f06eb373bf266f49c82bed20490f38a3850b038b88fc221de16" exitCode=0 Nov 24 08:43:28 crc kubenswrapper[4831]: I1124 08:43:28.694961 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" event={"ID":"23122f3c-2de2-4a17-b899-d71ea63b8b7f","Type":"ContainerDied","Data":"4ff51baa7b6f3f06eb373bf266f49c82bed20490f38a3850b038b88fc221de16"} Nov 24 08:43:28 crc kubenswrapper[4831]: I1124 08:43:28.894096 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:43:28 crc kubenswrapper[4831]: E1124 08:43:28.894932 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.053178 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.209980 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory\") pod \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.210574 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vtqg\" (UniqueName: \"kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg\") pod \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.210780 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key\") pod \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\" (UID: \"23122f3c-2de2-4a17-b899-d71ea63b8b7f\") " Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.221661 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg" (OuterVolumeSpecName: "kube-api-access-9vtqg") pod "23122f3c-2de2-4a17-b899-d71ea63b8b7f" (UID: "23122f3c-2de2-4a17-b899-d71ea63b8b7f"). InnerVolumeSpecName "kube-api-access-9vtqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.237467 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory" (OuterVolumeSpecName: "inventory") pod "23122f3c-2de2-4a17-b899-d71ea63b8b7f" (UID: "23122f3c-2de2-4a17-b899-d71ea63b8b7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.241156 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "23122f3c-2de2-4a17-b899-d71ea63b8b7f" (UID: "23122f3c-2de2-4a17-b899-d71ea63b8b7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.314005 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vtqg\" (UniqueName: \"kubernetes.io/projected/23122f3c-2de2-4a17-b899-d71ea63b8b7f-kube-api-access-9vtqg\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.314237 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.314555 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/23122f3c-2de2-4a17-b899-d71ea63b8b7f-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.717550 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" event={"ID":"23122f3c-2de2-4a17-b899-d71ea63b8b7f","Type":"ContainerDied","Data":"c5ee9137f76b460522afb9fac979e6202095ba0d25fb3b353c037ddf338c83b3"} Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.717603 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5ee9137f76b460522afb9fac979e6202095ba0d25fb3b353c037ddf338c83b3" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.717690 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jl4hn" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.824890 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq"] Nov 24 08:43:30 crc kubenswrapper[4831]: E1124 08:43:30.825547 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23122f3c-2de2-4a17-b899-d71ea63b8b7f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.825657 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="23122f3c-2de2-4a17-b899-d71ea63b8b7f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.825953 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="23122f3c-2de2-4a17-b899-d71ea63b8b7f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.826765 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.829340 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.830944 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.840912 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq"] Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.842068 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:43:30 crc kubenswrapper[4831]: I1124 08:43:30.842417 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.027874 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.028409 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.029207 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrrdf\" (UniqueName: \"kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.132190 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.132248 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.132311 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrrdf\" (UniqueName: \"kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.137961 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.138349 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.167607 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrrdf\" (UniqueName: \"kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.199678 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:31 crc kubenswrapper[4831]: I1124 08:43:31.774061 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq"] Nov 24 08:43:32 crc kubenswrapper[4831]: I1124 08:43:32.746955 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" event={"ID":"8200ad7a-4b9e-46f0-861b-1fc331d347e3","Type":"ContainerStarted","Data":"3d31a851272d263e4cac1d1c120b3d6f418a5f1593ba1256bf861a866160d941"} Nov 24 08:43:32 crc kubenswrapper[4831]: I1124 08:43:32.747023 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" event={"ID":"8200ad7a-4b9e-46f0-861b-1fc331d347e3","Type":"ContainerStarted","Data":"60b9ea9b6cdecab980845a22920253357c73b3e2ebabf3e87f496e7cc80179dd"} Nov 24 08:43:32 crc kubenswrapper[4831]: I1124 08:43:32.767776 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" podStartSLOduration=2.368588338 podStartE2EDuration="2.767759807s" podCreationTimestamp="2025-11-24 08:43:30 +0000 UTC" firstStartedPulling="2025-11-24 08:43:31.760008847 +0000 UTC m=+1685.635154050" lastFinishedPulling="2025-11-24 08:43:32.159180346 +0000 UTC m=+1686.034325519" observedRunningTime="2025-11-24 08:43:32.766707127 +0000 UTC m=+1686.641852280" watchObservedRunningTime="2025-11-24 08:43:32.767759807 +0000 UTC m=+1686.642904960" Nov 24 08:43:36 crc kubenswrapper[4831]: I1124 08:43:36.786410 4831 generic.go:334] "Generic (PLEG): container finished" podID="8200ad7a-4b9e-46f0-861b-1fc331d347e3" containerID="3d31a851272d263e4cac1d1c120b3d6f418a5f1593ba1256bf861a866160d941" exitCode=0 Nov 24 08:43:36 crc kubenswrapper[4831]: I1124 08:43:36.786547 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" event={"ID":"8200ad7a-4b9e-46f0-861b-1fc331d347e3","Type":"ContainerDied","Data":"3d31a851272d263e4cac1d1c120b3d6f418a5f1593ba1256bf861a866160d941"} Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.169678 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.270657 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrrdf\" (UniqueName: \"kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf\") pod \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.270907 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory\") pod \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.270964 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key\") pod \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\" (UID: \"8200ad7a-4b9e-46f0-861b-1fc331d347e3\") " Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.284954 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf" (OuterVolumeSpecName: "kube-api-access-nrrdf") pod "8200ad7a-4b9e-46f0-861b-1fc331d347e3" (UID: "8200ad7a-4b9e-46f0-861b-1fc331d347e3"). InnerVolumeSpecName "kube-api-access-nrrdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.302533 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8200ad7a-4b9e-46f0-861b-1fc331d347e3" (UID: "8200ad7a-4b9e-46f0-861b-1fc331d347e3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.303175 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory" (OuterVolumeSpecName: "inventory") pod "8200ad7a-4b9e-46f0-861b-1fc331d347e3" (UID: "8200ad7a-4b9e-46f0-861b-1fc331d347e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.374293 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.374398 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8200ad7a-4b9e-46f0-861b-1fc331d347e3-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.374422 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrrdf\" (UniqueName: \"kubernetes.io/projected/8200ad7a-4b9e-46f0-861b-1fc331d347e3-kube-api-access-nrrdf\") on node \"crc\" DevicePath \"\"" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.803308 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" event={"ID":"8200ad7a-4b9e-46f0-861b-1fc331d347e3","Type":"ContainerDied","Data":"60b9ea9b6cdecab980845a22920253357c73b3e2ebabf3e87f496e7cc80179dd"} Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.803368 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.803375 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60b9ea9b6cdecab980845a22920253357c73b3e2ebabf3e87f496e7cc80179dd" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.873029 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm"] Nov 24 08:43:38 crc kubenswrapper[4831]: E1124 08:43:38.873478 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8200ad7a-4b9e-46f0-861b-1fc331d347e3" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.873503 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="8200ad7a-4b9e-46f0-861b-1fc331d347e3" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.873703 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="8200ad7a-4b9e-46f0-861b-1fc331d347e3" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.874411 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.876407 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.876596 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.876948 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.877123 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.884871 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn4br\" (UniqueName: \"kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.885015 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.885096 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.903403 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm"] Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.986223 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.986835 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.986899 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn4br\" (UniqueName: \"kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.991998 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:38 crc kubenswrapper[4831]: I1124 08:43:38.994964 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:39 crc kubenswrapper[4831]: I1124 08:43:39.008434 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn4br\" (UniqueName: \"kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:39 crc kubenswrapper[4831]: I1124 08:43:39.192950 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:43:40 crc kubenswrapper[4831]: I1124 08:43:39.707570 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm"] Nov 24 08:43:40 crc kubenswrapper[4831]: I1124 08:43:39.816059 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" event={"ID":"88cf07ba-1cec-473e-8306-9b1594f166f5","Type":"ContainerStarted","Data":"beeeed9457751b4644bf6b47c92cc779ee8b9f92c0912388fbaf33b5f4babd3e"} Nov 24 08:43:40 crc kubenswrapper[4831]: I1124 08:43:40.827196 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" event={"ID":"88cf07ba-1cec-473e-8306-9b1594f166f5","Type":"ContainerStarted","Data":"807f288283cb6b5d77cfdd81a6eea0414e9a806be2a4064599ac3f7f96f5ee00"} Nov 24 08:43:40 crc kubenswrapper[4831]: I1124 08:43:40.849473 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" podStartSLOduration=2.270769064 podStartE2EDuration="2.849426897s" podCreationTimestamp="2025-11-24 08:43:38 +0000 UTC" firstStartedPulling="2025-11-24 08:43:39.708268718 +0000 UTC m=+1693.583413871" lastFinishedPulling="2025-11-24 08:43:40.286926541 +0000 UTC m=+1694.162071704" observedRunningTime="2025-11-24 08:43:40.842440969 +0000 UTC m=+1694.717586122" watchObservedRunningTime="2025-11-24 08:43:40.849426897 +0000 UTC m=+1694.724572060" Nov 24 08:43:40 crc kubenswrapper[4831]: I1124 08:43:40.896117 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:43:40 crc kubenswrapper[4831]: E1124 08:43:40.897835 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:43:52 crc kubenswrapper[4831]: I1124 08:43:52.893807 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:43:52 crc kubenswrapper[4831]: E1124 08:43:52.894794 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:43:57 crc kubenswrapper[4831]: I1124 08:43:57.041643 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-8vhg7"] Nov 24 08:43:57 crc kubenswrapper[4831]: I1124 08:43:57.051832 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ed31-account-create-fw7hs"] Nov 24 08:43:57 crc kubenswrapper[4831]: I1124 08:43:57.064619 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-8vhg7"] Nov 24 08:43:57 crc kubenswrapper[4831]: I1124 08:43:57.071739 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ed31-account-create-fw7hs"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.032096 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-8f65-account-create-96zkh"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.058035 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-zdfb7"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.070437 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5z8s4"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.077749 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1900-account-create-z9bh8"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.083638 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1900-account-create-z9bh8"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.089632 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-zdfb7"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.095630 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5z8s4"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.101527 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-8f65-account-create-96zkh"] Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.905070 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="004ad589-f884-4789-a566-e2076559de28" path="/var/lib/kubelet/pods/004ad589-f884-4789-a566-e2076559de28/volumes" Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.905933 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe" path="/var/lib/kubelet/pods/2fb1dd7b-8f0b-4d59-82a8-58e5551b9cfe/volumes" Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.906481 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aee4d7a-3f47-4df4-ba1b-94af4e39a495" path="/var/lib/kubelet/pods/6aee4d7a-3f47-4df4-ba1b-94af4e39a495/volumes" Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.906974 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b4819b3-74f6-4015-a437-412e5f4dff2d" path="/var/lib/kubelet/pods/7b4819b3-74f6-4015-a437-412e5f4dff2d/volumes" Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.907858 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6e8cfc9-9958-4540-9b4b-3166796bdbd5" path="/var/lib/kubelet/pods/a6e8cfc9-9958-4540-9b4b-3166796bdbd5/volumes" Nov 24 08:43:58 crc kubenswrapper[4831]: I1124 08:43:58.908402 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b639f94c-6740-497d-b53e-4c297f65d750" path="/var/lib/kubelet/pods/b639f94c-6740-497d-b53e-4c297f65d750/volumes" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.646908 4831 scope.go:117] "RemoveContainer" containerID="25d33da447560425d95c099131214a7ee1e8e06c809f30e2ef70a745e517c03b" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.672607 4831 scope.go:117] "RemoveContainer" containerID="c438e80f8e8fb25efa46f5b24d875c673edaad273345855a8bcd82af2fa4b5f5" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.710880 4831 scope.go:117] "RemoveContainer" containerID="76d923d0c152a7bfda17f33fb79382221ba13750f2fe9b04e6fdf91a4ba9c901" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.749350 4831 scope.go:117] "RemoveContainer" containerID="45edcf2b7d1862896e4dfb5ab8621241c6f1e13ebd0f8b512a406a8810b79b59" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.785276 4831 scope.go:117] "RemoveContainer" containerID="2c1339dbdb125375189623fed665b0cf1f79e5068870958ef2c0c843ac1f510c" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.846474 4831 scope.go:117] "RemoveContainer" containerID="013f5a978118240d2d9d2e4e24bccd03590a49ec5061457ba587e7bd2f99cae0" Nov 24 08:44:03 crc kubenswrapper[4831]: I1124 08:44:03.884954 4831 scope.go:117] "RemoveContainer" containerID="3c8ed503732d4ec2e1bec2fe4d9787a75cc99ab96e5c41247be24af5896773a4" Nov 24 08:44:05 crc kubenswrapper[4831]: I1124 08:44:05.893922 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:44:05 crc kubenswrapper[4831]: E1124 08:44:05.894640 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:44:17 crc kubenswrapper[4831]: I1124 08:44:17.893403 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:44:17 crc kubenswrapper[4831]: E1124 08:44:17.894430 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:44:24 crc kubenswrapper[4831]: I1124 08:44:24.049424 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ckzq6"] Nov 24 08:44:24 crc kubenswrapper[4831]: I1124 08:44:24.068181 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ckzq6"] Nov 24 08:44:24 crc kubenswrapper[4831]: I1124 08:44:24.904867 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff4e0931-319b-4ee2-bbef-224d6393f94a" path="/var/lib/kubelet/pods/ff4e0931-319b-4ee2-bbef-224d6393f94a/volumes" Nov 24 08:44:28 crc kubenswrapper[4831]: I1124 08:44:28.893153 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:44:28 crc kubenswrapper[4831]: E1124 08:44:28.894980 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:44:33 crc kubenswrapper[4831]: I1124 08:44:33.258304 4831 generic.go:334] "Generic (PLEG): container finished" podID="88cf07ba-1cec-473e-8306-9b1594f166f5" containerID="807f288283cb6b5d77cfdd81a6eea0414e9a806be2a4064599ac3f7f96f5ee00" exitCode=0 Nov 24 08:44:33 crc kubenswrapper[4831]: I1124 08:44:33.258444 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" event={"ID":"88cf07ba-1cec-473e-8306-9b1594f166f5","Type":"ContainerDied","Data":"807f288283cb6b5d77cfdd81a6eea0414e9a806be2a4064599ac3f7f96f5ee00"} Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.675557 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.814734 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn4br\" (UniqueName: \"kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br\") pod \"88cf07ba-1cec-473e-8306-9b1594f166f5\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.814838 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory\") pod \"88cf07ba-1cec-473e-8306-9b1594f166f5\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.814904 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key\") pod \"88cf07ba-1cec-473e-8306-9b1594f166f5\" (UID: \"88cf07ba-1cec-473e-8306-9b1594f166f5\") " Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.820689 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br" (OuterVolumeSpecName: "kube-api-access-qn4br") pod "88cf07ba-1cec-473e-8306-9b1594f166f5" (UID: "88cf07ba-1cec-473e-8306-9b1594f166f5"). InnerVolumeSpecName "kube-api-access-qn4br". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.842773 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory" (OuterVolumeSpecName: "inventory") pod "88cf07ba-1cec-473e-8306-9b1594f166f5" (UID: "88cf07ba-1cec-473e-8306-9b1594f166f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.843383 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88cf07ba-1cec-473e-8306-9b1594f166f5" (UID: "88cf07ba-1cec-473e-8306-9b1594f166f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.917193 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn4br\" (UniqueName: \"kubernetes.io/projected/88cf07ba-1cec-473e-8306-9b1594f166f5-kube-api-access-qn4br\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.917232 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:34 crc kubenswrapper[4831]: I1124 08:44:34.917241 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88cf07ba-1cec-473e-8306-9b1594f166f5-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.278679 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" event={"ID":"88cf07ba-1cec-473e-8306-9b1594f166f5","Type":"ContainerDied","Data":"beeeed9457751b4644bf6b47c92cc779ee8b9f92c0912388fbaf33b5f4babd3e"} Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.278731 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beeeed9457751b4644bf6b47c92cc779ee8b9f92c0912388fbaf33b5f4babd3e" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.279031 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.370229 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qbqmz"] Nov 24 08:44:35 crc kubenswrapper[4831]: E1124 08:44:35.372417 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88cf07ba-1cec-473e-8306-9b1594f166f5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.372454 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="88cf07ba-1cec-473e-8306-9b1594f166f5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.372688 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="88cf07ba-1cec-473e-8306-9b1594f166f5" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.373286 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.378750 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.381915 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.382185 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.382397 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.388116 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qbqmz"] Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.527705 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.527784 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.527869 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m9hk\" (UniqueName: \"kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.629972 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.630034 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.630088 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m9hk\" (UniqueName: \"kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.633886 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.637756 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.653438 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m9hk\" (UniqueName: \"kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk\") pod \"ssh-known-hosts-edpm-deployment-qbqmz\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:35 crc kubenswrapper[4831]: I1124 08:44:35.696006 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:36 crc kubenswrapper[4831]: W1124 08:44:36.224673 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fe18501_edce_4461_ab81_04b861259785.slice/crio-c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c WatchSource:0}: Error finding container c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c: Status 404 returned error can't find the container with id c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c Nov 24 08:44:36 crc kubenswrapper[4831]: I1124 08:44:36.227296 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-qbqmz"] Nov 24 08:44:36 crc kubenswrapper[4831]: I1124 08:44:36.287646 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" event={"ID":"5fe18501-edce-4461-ab81-04b861259785","Type":"ContainerStarted","Data":"c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c"} Nov 24 08:44:37 crc kubenswrapper[4831]: I1124 08:44:37.298971 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" event={"ID":"5fe18501-edce-4461-ab81-04b861259785","Type":"ContainerStarted","Data":"e27c7f03c401c970c7c0cd18342c20cdc0d9b5c80f443ba02e7eb9a6b6fdf882"} Nov 24 08:44:37 crc kubenswrapper[4831]: I1124 08:44:37.323170 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" podStartSLOduration=1.95231787 podStartE2EDuration="2.323146786s" podCreationTimestamp="2025-11-24 08:44:35 +0000 UTC" firstStartedPulling="2025-11-24 08:44:36.234932438 +0000 UTC m=+1750.110077591" lastFinishedPulling="2025-11-24 08:44:36.605761304 +0000 UTC m=+1750.480906507" observedRunningTime="2025-11-24 08:44:37.315010015 +0000 UTC m=+1751.190155168" watchObservedRunningTime="2025-11-24 08:44:37.323146786 +0000 UTC m=+1751.198291929" Nov 24 08:44:40 crc kubenswrapper[4831]: I1124 08:44:40.903613 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:44:40 crc kubenswrapper[4831]: E1124 08:44:40.909139 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:44:44 crc kubenswrapper[4831]: I1124 08:44:44.037268 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhjsw"] Nov 24 08:44:44 crc kubenswrapper[4831]: I1124 08:44:44.048088 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lhjsw"] Nov 24 08:44:44 crc kubenswrapper[4831]: I1124 08:44:44.358809 4831 generic.go:334] "Generic (PLEG): container finished" podID="5fe18501-edce-4461-ab81-04b861259785" containerID="e27c7f03c401c970c7c0cd18342c20cdc0d9b5c80f443ba02e7eb9a6b6fdf882" exitCode=0 Nov 24 08:44:44 crc kubenswrapper[4831]: I1124 08:44:44.358972 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" event={"ID":"5fe18501-edce-4461-ab81-04b861259785","Type":"ContainerDied","Data":"e27c7f03c401c970c7c0cd18342c20cdc0d9b5c80f443ba02e7eb9a6b6fdf882"} Nov 24 08:44:44 crc kubenswrapper[4831]: I1124 08:44:44.906449 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fbc0462-5e34-49ff-8786-5806fbbf36ec" path="/var/lib/kubelet/pods/8fbc0462-5e34-49ff-8786-5806fbbf36ec/volumes" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.720139 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.845881 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m9hk\" (UniqueName: \"kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk\") pod \"5fe18501-edce-4461-ab81-04b861259785\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.846097 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0\") pod \"5fe18501-edce-4461-ab81-04b861259785\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.846224 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam\") pod \"5fe18501-edce-4461-ab81-04b861259785\" (UID: \"5fe18501-edce-4461-ab81-04b861259785\") " Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.851522 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk" (OuterVolumeSpecName: "kube-api-access-6m9hk") pod "5fe18501-edce-4461-ab81-04b861259785" (UID: "5fe18501-edce-4461-ab81-04b861259785"). InnerVolumeSpecName "kube-api-access-6m9hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.870631 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "5fe18501-edce-4461-ab81-04b861259785" (UID: "5fe18501-edce-4461-ab81-04b861259785"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.874615 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5fe18501-edce-4461-ab81-04b861259785" (UID: "5fe18501-edce-4461-ab81-04b861259785"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.948768 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m9hk\" (UniqueName: \"kubernetes.io/projected/5fe18501-edce-4461-ab81-04b861259785-kube-api-access-6m9hk\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.948807 4831 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:45 crc kubenswrapper[4831]: I1124 08:44:45.948818 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5fe18501-edce-4461-ab81-04b861259785-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.037216 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vl2xc"] Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.044362 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vl2xc"] Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.377209 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" event={"ID":"5fe18501-edce-4461-ab81-04b861259785","Type":"ContainerDied","Data":"c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c"} Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.377245 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-qbqmz" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.377256 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1741a2ede738418072519051ef8d77ed529494f09d7f1a3c19aeb6ee71feb1c" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.549752 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9"] Nov 24 08:44:46 crc kubenswrapper[4831]: E1124 08:44:46.550464 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe18501-edce-4461-ab81-04b861259785" containerName="ssh-known-hosts-edpm-deployment" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.550481 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe18501-edce-4461-ab81-04b861259785" containerName="ssh-known-hosts-edpm-deployment" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.550633 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe18501-edce-4461-ab81-04b861259785" containerName="ssh-known-hosts-edpm-deployment" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.551233 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.554229 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.554503 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.555351 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.556361 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.571396 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9"] Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.661465 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz4db\" (UniqueName: \"kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.661531 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.661833 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.764103 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.764235 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz4db\" (UniqueName: \"kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.764287 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.768778 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.783953 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.785666 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz4db\" (UniqueName: \"kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zmdx9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.872518 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:46 crc kubenswrapper[4831]: I1124 08:44:46.911979 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd" path="/var/lib/kubelet/pods/2f1182a5-f1f1-4658-9f1a-b4cd3e9f1cfd/volumes" Nov 24 08:44:47 crc kubenswrapper[4831]: I1124 08:44:47.465231 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9"] Nov 24 08:44:47 crc kubenswrapper[4831]: W1124 08:44:47.469029 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa8c7ab9_41a6_412f_9abf_bc4105b768d9.slice/crio-a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9 WatchSource:0}: Error finding container a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9: Status 404 returned error can't find the container with id a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9 Nov 24 08:44:48 crc kubenswrapper[4831]: I1124 08:44:48.395152 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" event={"ID":"aa8c7ab9-41a6-412f-9abf-bc4105b768d9","Type":"ContainerStarted","Data":"fc00dbffa3fa7e3062e4da9112999e042f4c2e912df351c0576c801f9e0f7fcf"} Nov 24 08:44:48 crc kubenswrapper[4831]: I1124 08:44:48.396441 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" event={"ID":"aa8c7ab9-41a6-412f-9abf-bc4105b768d9","Type":"ContainerStarted","Data":"a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9"} Nov 24 08:44:48 crc kubenswrapper[4831]: I1124 08:44:48.426917 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" podStartSLOduration=1.972007673 podStartE2EDuration="2.426894638s" podCreationTimestamp="2025-11-24 08:44:46 +0000 UTC" firstStartedPulling="2025-11-24 08:44:47.472340634 +0000 UTC m=+1761.347485787" lastFinishedPulling="2025-11-24 08:44:47.927227599 +0000 UTC m=+1761.802372752" observedRunningTime="2025-11-24 08:44:48.421748361 +0000 UTC m=+1762.296893514" watchObservedRunningTime="2025-11-24 08:44:48.426894638 +0000 UTC m=+1762.302039811" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.547625 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.549910 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.567895 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.737287 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brvcf\" (UniqueName: \"kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.737370 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.737591 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.839151 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.839258 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brvcf\" (UniqueName: \"kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.839311 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.839916 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.839963 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.863634 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brvcf\" (UniqueName: \"kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf\") pod \"community-operators-sb4dk\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:50 crc kubenswrapper[4831]: I1124 08:44:50.877162 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:44:51 crc kubenswrapper[4831]: I1124 08:44:51.379587 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:44:51 crc kubenswrapper[4831]: I1124 08:44:51.418559 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerStarted","Data":"be663b8f88595cdf62e082f668b45a92a7a674e19b60373428a0e2c7c8286fdf"} Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.342114 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.344415 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.356109 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.435196 4831 generic.go:334] "Generic (PLEG): container finished" podID="338af098-27e8-43ca-aa83-62c258e735d5" containerID="c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d" exitCode=0 Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.435266 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerDied","Data":"c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d"} Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.468604 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.468768 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.468806 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rrcv\" (UniqueName: \"kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.570059 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rrcv\" (UniqueName: \"kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.570197 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.570307 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.571042 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.571208 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.591262 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rrcv\" (UniqueName: \"kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv\") pod \"redhat-marketplace-6hpkr\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.679270 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:44:52 crc kubenswrapper[4831]: I1124 08:44:52.896521 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:44:52 crc kubenswrapper[4831]: E1124 08:44:52.897199 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:44:53 crc kubenswrapper[4831]: I1124 08:44:53.181840 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:44:53 crc kubenswrapper[4831]: I1124 08:44:53.446172 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerStarted","Data":"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea"} Nov 24 08:44:53 crc kubenswrapper[4831]: I1124 08:44:53.448258 4831 generic.go:334] "Generic (PLEG): container finished" podID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerID="0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67" exitCode=0 Nov 24 08:44:53 crc kubenswrapper[4831]: I1124 08:44:53.448426 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerDied","Data":"0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67"} Nov 24 08:44:53 crc kubenswrapper[4831]: I1124 08:44:53.448464 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerStarted","Data":"a6cd653b4771db86361a4e3bbe33463149b279efad95163aaa09540f5843d2a7"} Nov 24 08:44:54 crc kubenswrapper[4831]: I1124 08:44:54.460111 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerStarted","Data":"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2"} Nov 24 08:44:55 crc kubenswrapper[4831]: I1124 08:44:55.468583 4831 generic.go:334] "Generic (PLEG): container finished" podID="338af098-27e8-43ca-aa83-62c258e735d5" containerID="ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea" exitCode=0 Nov 24 08:44:55 crc kubenswrapper[4831]: I1124 08:44:55.468758 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerDied","Data":"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea"} Nov 24 08:44:56 crc kubenswrapper[4831]: I1124 08:44:56.480850 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerStarted","Data":"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8"} Nov 24 08:44:56 crc kubenswrapper[4831]: I1124 08:44:56.483252 4831 generic.go:334] "Generic (PLEG): container finished" podID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerID="3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2" exitCode=0 Nov 24 08:44:56 crc kubenswrapper[4831]: I1124 08:44:56.483422 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerDied","Data":"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2"} Nov 24 08:44:56 crc kubenswrapper[4831]: I1124 08:44:56.537135 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sb4dk" podStartSLOduration=3.125688826 podStartE2EDuration="6.537103914s" podCreationTimestamp="2025-11-24 08:44:50 +0000 UTC" firstStartedPulling="2025-11-24 08:44:52.439556724 +0000 UTC m=+1766.314701877" lastFinishedPulling="2025-11-24 08:44:55.850971802 +0000 UTC m=+1769.726116965" observedRunningTime="2025-11-24 08:44:56.510721224 +0000 UTC m=+1770.385866427" watchObservedRunningTime="2025-11-24 08:44:56.537103914 +0000 UTC m=+1770.412249097" Nov 24 08:44:57 crc kubenswrapper[4831]: I1124 08:44:57.493740 4831 generic.go:334] "Generic (PLEG): container finished" podID="aa8c7ab9-41a6-412f-9abf-bc4105b768d9" containerID="fc00dbffa3fa7e3062e4da9112999e042f4c2e912df351c0576c801f9e0f7fcf" exitCode=0 Nov 24 08:44:57 crc kubenswrapper[4831]: I1124 08:44:57.493867 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" event={"ID":"aa8c7ab9-41a6-412f-9abf-bc4105b768d9","Type":"ContainerDied","Data":"fc00dbffa3fa7e3062e4da9112999e042f4c2e912df351c0576c801f9e0f7fcf"} Nov 24 08:44:58 crc kubenswrapper[4831]: I1124 08:44:58.504677 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerStarted","Data":"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486"} Nov 24 08:44:58 crc kubenswrapper[4831]: I1124 08:44:58.531120 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6hpkr" podStartSLOduration=2.859501988 podStartE2EDuration="6.531099504s" podCreationTimestamp="2025-11-24 08:44:52 +0000 UTC" firstStartedPulling="2025-11-24 08:44:53.450609705 +0000 UTC m=+1767.325754858" lastFinishedPulling="2025-11-24 08:44:57.122207211 +0000 UTC m=+1770.997352374" observedRunningTime="2025-11-24 08:44:58.526217075 +0000 UTC m=+1772.401362248" watchObservedRunningTime="2025-11-24 08:44:58.531099504 +0000 UTC m=+1772.406244657" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.045059 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.108525 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory\") pod \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.108677 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz4db\" (UniqueName: \"kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db\") pod \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.108964 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key\") pod \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\" (UID: \"aa8c7ab9-41a6-412f-9abf-bc4105b768d9\") " Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.124518 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db" (OuterVolumeSpecName: "kube-api-access-pz4db") pod "aa8c7ab9-41a6-412f-9abf-bc4105b768d9" (UID: "aa8c7ab9-41a6-412f-9abf-bc4105b768d9"). InnerVolumeSpecName "kube-api-access-pz4db". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.149047 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory" (OuterVolumeSpecName: "inventory") pod "aa8c7ab9-41a6-412f-9abf-bc4105b768d9" (UID: "aa8c7ab9-41a6-412f-9abf-bc4105b768d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.190550 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aa8c7ab9-41a6-412f-9abf-bc4105b768d9" (UID: "aa8c7ab9-41a6-412f-9abf-bc4105b768d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.211703 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.211769 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz4db\" (UniqueName: \"kubernetes.io/projected/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-kube-api-access-pz4db\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.211783 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aa8c7ab9-41a6-412f-9abf-bc4105b768d9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.514568 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" event={"ID":"aa8c7ab9-41a6-412f-9abf-bc4105b768d9","Type":"ContainerDied","Data":"a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9"} Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.514605 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8f1b7bd1c660e26d5661f558265beeb1c2c3d5beff6c17a3425aef3500191e9" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.514615 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zmdx9" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.584529 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d"] Nov 24 08:44:59 crc kubenswrapper[4831]: E1124 08:44:59.585399 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8c7ab9-41a6-412f-9abf-bc4105b768d9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.585423 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8c7ab9-41a6-412f-9abf-bc4105b768d9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.585664 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa8c7ab9-41a6-412f-9abf-bc4105b768d9" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.587037 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.589658 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.591107 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.591483 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gmfqp" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.591632 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.617806 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfjqj\" (UniqueName: \"kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.618163 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.618310 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.624148 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d"] Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.720158 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.720224 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.720304 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfjqj\" (UniqueName: \"kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.724014 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.724014 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.736910 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfjqj\" (UniqueName: \"kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:44:59 crc kubenswrapper[4831]: I1124 08:44:59.932496 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.145806 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs"] Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.147107 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.149905 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.150090 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.164389 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs"] Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.229291 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.229874 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvj7t\" (UniqueName: \"kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.229956 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.331738 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvj7t\" (UniqueName: \"kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.331826 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.333066 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.333834 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.340754 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.356209 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvj7t\" (UniqueName: \"kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t\") pod \"collect-profiles-29399565-dfncs\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.457596 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d"] Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.482106 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.522810 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" event={"ID":"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7","Type":"ContainerStarted","Data":"73317cbc7b66c6df683862ade5914dd1dd4fcfbdce0b18a646d24308966033c0"} Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.878717 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.879064 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:00 crc kubenswrapper[4831]: I1124 08:45:00.941077 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs"] Nov 24 08:45:00 crc kubenswrapper[4831]: W1124 08:45:00.955304 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab136ed_1e42_4e4f_9eea_a1268dc13add.slice/crio-7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34 WatchSource:0}: Error finding container 7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34: Status 404 returned error can't find the container with id 7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34 Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.533480 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" event={"ID":"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7","Type":"ContainerStarted","Data":"82061b2d1c2824ac7d09697cf690f1a9ed2015cf723a0fe7acdcf6ff0566a8f5"} Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.538889 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" event={"ID":"8ab136ed-1e42-4e4f-9eea-a1268dc13add","Type":"ContainerStarted","Data":"cdddb238f775634733aaaa2482e08ed2405f228ae49ff047857a971894df8786"} Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.538925 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" event={"ID":"8ab136ed-1e42-4e4f-9eea-a1268dc13add","Type":"ContainerStarted","Data":"7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34"} Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.556349 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" podStartSLOduration=2.103300328 podStartE2EDuration="2.556326791s" podCreationTimestamp="2025-11-24 08:44:59 +0000 UTC" firstStartedPulling="2025-11-24 08:45:00.468331912 +0000 UTC m=+1774.343477065" lastFinishedPulling="2025-11-24 08:45:00.921358375 +0000 UTC m=+1774.796503528" observedRunningTime="2025-11-24 08:45:01.550036013 +0000 UTC m=+1775.425181166" watchObservedRunningTime="2025-11-24 08:45:01.556326791 +0000 UTC m=+1775.431471944" Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.581311 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" podStartSLOduration=1.581295431 podStartE2EDuration="1.581295431s" podCreationTimestamp="2025-11-24 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 08:45:01.575776555 +0000 UTC m=+1775.450921718" watchObservedRunningTime="2025-11-24 08:45:01.581295431 +0000 UTC m=+1775.456440584" Nov 24 08:45:01 crc kubenswrapper[4831]: I1124 08:45:01.954337 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-sb4dk" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="registry-server" probeResult="failure" output=< Nov 24 08:45:01 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:45:01 crc kubenswrapper[4831]: > Nov 24 08:45:02 crc kubenswrapper[4831]: I1124 08:45:02.549610 4831 generic.go:334] "Generic (PLEG): container finished" podID="8ab136ed-1e42-4e4f-9eea-a1268dc13add" containerID="cdddb238f775634733aaaa2482e08ed2405f228ae49ff047857a971894df8786" exitCode=0 Nov 24 08:45:02 crc kubenswrapper[4831]: I1124 08:45:02.549714 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" event={"ID":"8ab136ed-1e42-4e4f-9eea-a1268dc13add","Type":"ContainerDied","Data":"cdddb238f775634733aaaa2482e08ed2405f228ae49ff047857a971894df8786"} Nov 24 08:45:02 crc kubenswrapper[4831]: I1124 08:45:02.679926 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:02 crc kubenswrapper[4831]: I1124 08:45:02.682034 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:02 crc kubenswrapper[4831]: I1124 08:45:02.734585 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:03 crc kubenswrapper[4831]: I1124 08:45:03.617258 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:03 crc kubenswrapper[4831]: I1124 08:45:03.908372 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.000950 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume\") pod \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.001066 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume\") pod \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.001160 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvj7t\" (UniqueName: \"kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t\") pod \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\" (UID: \"8ab136ed-1e42-4e4f-9eea-a1268dc13add\") " Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.001738 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume" (OuterVolumeSpecName: "config-volume") pod "8ab136ed-1e42-4e4f-9eea-a1268dc13add" (UID: "8ab136ed-1e42-4e4f-9eea-a1268dc13add"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.007704 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t" (OuterVolumeSpecName: "kube-api-access-kvj7t") pod "8ab136ed-1e42-4e4f-9eea-a1268dc13add" (UID: "8ab136ed-1e42-4e4f-9eea-a1268dc13add"). InnerVolumeSpecName "kube-api-access-kvj7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.009394 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8ab136ed-1e42-4e4f-9eea-a1268dc13add" (UID: "8ab136ed-1e42-4e4f-9eea-a1268dc13add"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.025581 4831 scope.go:117] "RemoveContainer" containerID="847c5e0f764c84d0674b9661873594c47402a7b00fbe858162979eae881f5ac9" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.103273 4831 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ab136ed-1e42-4e4f-9eea-a1268dc13add-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.103342 4831 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ab136ed-1e42-4e4f-9eea-a1268dc13add-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.103353 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvj7t\" (UniqueName: \"kubernetes.io/projected/8ab136ed-1e42-4e4f-9eea-a1268dc13add-kube-api-access-kvj7t\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.112553 4831 scope.go:117] "RemoveContainer" containerID="e64e5de0eb9f0070417fe8b2a65d406a7511d5e8d42016b1916a01e64cff2573" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.162378 4831 scope.go:117] "RemoveContainer" containerID="e578b78210830b2e0a840fed543cf9da8fcdbf4715c4a33306b37635b129c57f" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.568870 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.568918 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399565-dfncs" event={"ID":"8ab136ed-1e42-4e4f-9eea-a1268dc13add","Type":"ContainerDied","Data":"7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34"} Nov 24 08:45:04 crc kubenswrapper[4831]: I1124 08:45:04.568995 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a9220662ed408e75467ed88db4ea01fe994e7c1e834a6770736ece28d0b6d34" Nov 24 08:45:05 crc kubenswrapper[4831]: I1124 08:45:05.736001 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:45:06 crc kubenswrapper[4831]: I1124 08:45:06.583404 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6hpkr" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="registry-server" containerID="cri-o://524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486" gracePeriod=2 Nov 24 08:45:06 crc kubenswrapper[4831]: I1124 08:45:06.899704 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:45:06 crc kubenswrapper[4831]: E1124 08:45:06.900406 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.008928 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.055036 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rrcv\" (UniqueName: \"kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv\") pod \"4f053270-a360-4893-813e-b8a8fe12c6a5\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.055120 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities\") pod \"4f053270-a360-4893-813e-b8a8fe12c6a5\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.055211 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content\") pod \"4f053270-a360-4893-813e-b8a8fe12c6a5\" (UID: \"4f053270-a360-4893-813e-b8a8fe12c6a5\") " Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.056613 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities" (OuterVolumeSpecName: "utilities") pod "4f053270-a360-4893-813e-b8a8fe12c6a5" (UID: "4f053270-a360-4893-813e-b8a8fe12c6a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.062417 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv" (OuterVolumeSpecName: "kube-api-access-9rrcv") pod "4f053270-a360-4893-813e-b8a8fe12c6a5" (UID: "4f053270-a360-4893-813e-b8a8fe12c6a5"). InnerVolumeSpecName "kube-api-access-9rrcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.076730 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f053270-a360-4893-813e-b8a8fe12c6a5" (UID: "4f053270-a360-4893-813e-b8a8fe12c6a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.157146 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rrcv\" (UniqueName: \"kubernetes.io/projected/4f053270-a360-4893-813e-b8a8fe12c6a5-kube-api-access-9rrcv\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.157185 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.157198 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f053270-a360-4893-813e-b8a8fe12c6a5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.599642 4831 generic.go:334] "Generic (PLEG): container finished" podID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerID="524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486" exitCode=0 Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.599684 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerDied","Data":"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486"} Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.599712 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hpkr" event={"ID":"4f053270-a360-4893-813e-b8a8fe12c6a5","Type":"ContainerDied","Data":"a6cd653b4771db86361a4e3bbe33463149b279efad95163aaa09540f5843d2a7"} Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.599731 4831 scope.go:117] "RemoveContainer" containerID="524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.599736 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hpkr" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.628918 4831 scope.go:117] "RemoveContainer" containerID="3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.652086 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.657713 4831 scope.go:117] "RemoveContainer" containerID="0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.662673 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hpkr"] Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.686413 4831 scope.go:117] "RemoveContainer" containerID="524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486" Nov 24 08:45:07 crc kubenswrapper[4831]: E1124 08:45:07.687047 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486\": container with ID starting with 524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486 not found: ID does not exist" containerID="524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.687078 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486"} err="failed to get container status \"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486\": rpc error: code = NotFound desc = could not find container \"524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486\": container with ID starting with 524061524bb36740ce0d3b3a5df108d54813d2f1b2be223e04b8670988ad8486 not found: ID does not exist" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.687103 4831 scope.go:117] "RemoveContainer" containerID="3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2" Nov 24 08:45:07 crc kubenswrapper[4831]: E1124 08:45:07.688614 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2\": container with ID starting with 3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2 not found: ID does not exist" containerID="3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.688668 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2"} err="failed to get container status \"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2\": rpc error: code = NotFound desc = could not find container \"3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2\": container with ID starting with 3c240e761be66a9ea79e3de57969aa2f7699e253e7aebe83c30069360ed502c2 not found: ID does not exist" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.688690 4831 scope.go:117] "RemoveContainer" containerID="0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67" Nov 24 08:45:07 crc kubenswrapper[4831]: E1124 08:45:07.689623 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67\": container with ID starting with 0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67 not found: ID does not exist" containerID="0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67" Nov 24 08:45:07 crc kubenswrapper[4831]: I1124 08:45:07.689652 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67"} err="failed to get container status \"0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67\": rpc error: code = NotFound desc = could not find container \"0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67\": container with ID starting with 0573f25ba4752c2d8e9a41dc1ad78953bd25acf6574863f7febd34f6f50a7c67 not found: ID does not exist" Nov 24 08:45:08 crc kubenswrapper[4831]: I1124 08:45:08.903507 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" path="/var/lib/kubelet/pods/4f053270-a360-4893-813e-b8a8fe12c6a5/volumes" Nov 24 08:45:10 crc kubenswrapper[4831]: I1124 08:45:10.938052 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:10 crc kubenswrapper[4831]: I1124 08:45:10.997017 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:11 crc kubenswrapper[4831]: I1124 08:45:11.541789 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:45:11 crc kubenswrapper[4831]: I1124 08:45:11.636831 4831 generic.go:334] "Generic (PLEG): container finished" podID="95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" containerID="82061b2d1c2824ac7d09697cf690f1a9ed2015cf723a0fe7acdcf6ff0566a8f5" exitCode=0 Nov 24 08:45:11 crc kubenswrapper[4831]: I1124 08:45:11.636940 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" event={"ID":"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7","Type":"ContainerDied","Data":"82061b2d1c2824ac7d09697cf690f1a9ed2015cf723a0fe7acdcf6ff0566a8f5"} Nov 24 08:45:12 crc kubenswrapper[4831]: I1124 08:45:12.646390 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sb4dk" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="registry-server" containerID="cri-o://d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8" gracePeriod=2 Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.151949 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.161751 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178265 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key\") pod \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178523 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brvcf\" (UniqueName: \"kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf\") pod \"338af098-27e8-43ca-aa83-62c258e735d5\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178591 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content\") pod \"338af098-27e8-43ca-aa83-62c258e735d5\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178638 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities\") pod \"338af098-27e8-43ca-aa83-62c258e735d5\" (UID: \"338af098-27e8-43ca-aa83-62c258e735d5\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178675 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfjqj\" (UniqueName: \"kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj\") pod \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.178701 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory\") pod \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\" (UID: \"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7\") " Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.179463 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities" (OuterVolumeSpecName: "utilities") pod "338af098-27e8-43ca-aa83-62c258e735d5" (UID: "338af098-27e8-43ca-aa83-62c258e735d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.186192 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf" (OuterVolumeSpecName: "kube-api-access-brvcf") pod "338af098-27e8-43ca-aa83-62c258e735d5" (UID: "338af098-27e8-43ca-aa83-62c258e735d5"). InnerVolumeSpecName "kube-api-access-brvcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.195871 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj" (OuterVolumeSpecName: "kube-api-access-lfjqj") pod "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" (UID: "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7"). InnerVolumeSpecName "kube-api-access-lfjqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.218705 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" (UID: "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.241504 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory" (OuterVolumeSpecName: "inventory") pod "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" (UID: "95fe9895-4e23-4cb9-9eb1-1a76a94a31d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.254937 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "338af098-27e8-43ca-aa83-62c258e735d5" (UID: "338af098-27e8-43ca-aa83-62c258e735d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280899 4831 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280931 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brvcf\" (UniqueName: \"kubernetes.io/projected/338af098-27e8-43ca-aa83-62c258e735d5-kube-api-access-brvcf\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280943 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280971 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/338af098-27e8-43ca-aa83-62c258e735d5-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280981 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfjqj\" (UniqueName: \"kubernetes.io/projected/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-kube-api-access-lfjqj\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.280990 4831 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95fe9895-4e23-4cb9-9eb1-1a76a94a31d7-inventory\") on node \"crc\" DevicePath \"\"" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.657904 4831 generic.go:334] "Generic (PLEG): container finished" podID="338af098-27e8-43ca-aa83-62c258e735d5" containerID="d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8" exitCode=0 Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.658000 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerDied","Data":"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8"} Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.658039 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sb4dk" event={"ID":"338af098-27e8-43ca-aa83-62c258e735d5","Type":"ContainerDied","Data":"be663b8f88595cdf62e082f668b45a92a7a674e19b60373428a0e2c7c8286fdf"} Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.658041 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sb4dk" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.658064 4831 scope.go:117] "RemoveContainer" containerID="d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.662878 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" event={"ID":"95fe9895-4e23-4cb9-9eb1-1a76a94a31d7","Type":"ContainerDied","Data":"73317cbc7b66c6df683862ade5914dd1dd4fcfbdce0b18a646d24308966033c0"} Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.662943 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73317cbc7b66c6df683862ade5914dd1dd4fcfbdce0b18a646d24308966033c0" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.662945 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.711197 4831 scope.go:117] "RemoveContainer" containerID="ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.730501 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.738489 4831 scope.go:117] "RemoveContainer" containerID="c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.741556 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sb4dk"] Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.770501 4831 scope.go:117] "RemoveContainer" containerID="d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8" Nov 24 08:45:13 crc kubenswrapper[4831]: E1124 08:45:13.771557 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8\": container with ID starting with d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8 not found: ID does not exist" containerID="d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.771600 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8"} err="failed to get container status \"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8\": rpc error: code = NotFound desc = could not find container \"d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8\": container with ID starting with d83e98932c72ae395a4acc8e7834a0b5038443a2bbd51c45e39e39994c24bea8 not found: ID does not exist" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.771633 4831 scope.go:117] "RemoveContainer" containerID="ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea" Nov 24 08:45:13 crc kubenswrapper[4831]: E1124 08:45:13.772099 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea\": container with ID starting with ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea not found: ID does not exist" containerID="ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.772122 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea"} err="failed to get container status \"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea\": rpc error: code = NotFound desc = could not find container \"ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea\": container with ID starting with ed1f95f1acf60b10bb945b916b55bde63e96bde43b7423e14c3228506ea83bea not found: ID does not exist" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.772135 4831 scope.go:117] "RemoveContainer" containerID="c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d" Nov 24 08:45:13 crc kubenswrapper[4831]: E1124 08:45:13.772409 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d\": container with ID starting with c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d not found: ID does not exist" containerID="c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d" Nov 24 08:45:13 crc kubenswrapper[4831]: I1124 08:45:13.772429 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d"} err="failed to get container status \"c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d\": rpc error: code = NotFound desc = could not find container \"c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d\": container with ID starting with c4efea503ea1268387088eb90657813ffbc46f3681c36e059981a4f5ed49705d not found: ID does not exist" Nov 24 08:45:14 crc kubenswrapper[4831]: I1124 08:45:14.906027 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="338af098-27e8-43ca-aa83-62c258e735d5" path="/var/lib/kubelet/pods/338af098-27e8-43ca-aa83-62c258e735d5/volumes" Nov 24 08:45:17 crc kubenswrapper[4831]: I1124 08:45:17.893302 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:45:17 crc kubenswrapper[4831]: E1124 08:45:17.893873 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:45:30 crc kubenswrapper[4831]: I1124 08:45:30.043666 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-22ldq"] Nov 24 08:45:30 crc kubenswrapper[4831]: I1124 08:45:30.063490 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-22ldq"] Nov 24 08:45:30 crc kubenswrapper[4831]: I1124 08:45:30.910669 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2497432-87c5-4ec8-8bd0-5a27e60e6c45" path="/var/lib/kubelet/pods/c2497432-87c5-4ec8-8bd0-5a27e60e6c45/volumes" Nov 24 08:45:32 crc kubenswrapper[4831]: I1124 08:45:32.894035 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:45:32 crc kubenswrapper[4831]: E1124 08:45:32.894986 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:45:43 crc kubenswrapper[4831]: I1124 08:45:43.894027 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:45:43 crc kubenswrapper[4831]: E1124 08:45:43.894492 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:45:57 crc kubenswrapper[4831]: I1124 08:45:57.894074 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:45:57 crc kubenswrapper[4831]: E1124 08:45:57.894752 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:46:04 crc kubenswrapper[4831]: I1124 08:46:04.261593 4831 scope.go:117] "RemoveContainer" containerID="5df5ec28960fdd9ac8045f4be2310b118b924023266fd3addf77d19160cfab20" Nov 24 08:46:10 crc kubenswrapper[4831]: I1124 08:46:10.893701 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:46:11 crc kubenswrapper[4831]: I1124 08:46:11.219241 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce"} Nov 24 08:48:28 crc kubenswrapper[4831]: I1124 08:48:28.401787 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:48:28 crc kubenswrapper[4831]: I1124 08:48:28.402794 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:48:58 crc kubenswrapper[4831]: I1124 08:48:58.401160 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:48:58 crc kubenswrapper[4831]: I1124 08:48:58.401880 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.432921 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.433988 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434006 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434022 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="extract-utilities" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434029 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="extract-utilities" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434045 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ab136ed-1e42-4e4f-9eea-a1268dc13add" containerName="collect-profiles" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434055 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ab136ed-1e42-4e4f-9eea-a1268dc13add" containerName="collect-profiles" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434069 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434078 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434090 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="extract-utilities" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434098 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="extract-utilities" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434110 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434117 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434130 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="extract-content" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434137 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="extract-content" Nov 24 08:49:06 crc kubenswrapper[4831]: E1124 08:49:06.434149 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="extract-content" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434158 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="extract-content" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434389 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ab136ed-1e42-4e4f-9eea-a1268dc13add" containerName="collect-profiles" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434401 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="338af098-27e8-43ca-aa83-62c258e735d5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434409 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f053270-a360-4893-813e-b8a8fe12c6a5" containerName="registry-server" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.434416 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fe9895-4e23-4cb9-9eb1-1a76a94a31d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.436001 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.443703 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zslmx\" (UniqueName: \"kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.443758 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.443902 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.457414 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.545775 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zslmx\" (UniqueName: \"kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.545842 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.545875 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.546503 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.546521 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.567851 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zslmx\" (UniqueName: \"kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx\") pod \"redhat-operators-zvbpc\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:06 crc kubenswrapper[4831]: I1124 08:49:06.759648 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:07 crc kubenswrapper[4831]: I1124 08:49:07.267651 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:07 crc kubenswrapper[4831]: I1124 08:49:07.924570 4831 generic.go:334] "Generic (PLEG): container finished" podID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerID="af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9" exitCode=0 Nov 24 08:49:07 crc kubenswrapper[4831]: I1124 08:49:07.924617 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerDied","Data":"af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9"} Nov 24 08:49:07 crc kubenswrapper[4831]: I1124 08:49:07.924927 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerStarted","Data":"1c29c523461afac7bcc9a0c9123312ca5e0467e3b47e53f260c8467454025fde"} Nov 24 08:49:07 crc kubenswrapper[4831]: I1124 08:49:07.926762 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:49:09 crc kubenswrapper[4831]: I1124 08:49:09.945498 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerStarted","Data":"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae"} Nov 24 08:49:16 crc kubenswrapper[4831]: I1124 08:49:16.004056 4831 generic.go:334] "Generic (PLEG): container finished" podID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerID="a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae" exitCode=0 Nov 24 08:49:16 crc kubenswrapper[4831]: I1124 08:49:16.004410 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerDied","Data":"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae"} Nov 24 08:49:17 crc kubenswrapper[4831]: I1124 08:49:17.020639 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerStarted","Data":"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224"} Nov 24 08:49:17 crc kubenswrapper[4831]: I1124 08:49:17.042416 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zvbpc" podStartSLOduration=2.54292447 podStartE2EDuration="11.04238593s" podCreationTimestamp="2025-11-24 08:49:06 +0000 UTC" firstStartedPulling="2025-11-24 08:49:07.926556846 +0000 UTC m=+2021.801701999" lastFinishedPulling="2025-11-24 08:49:16.426018306 +0000 UTC m=+2030.301163459" observedRunningTime="2025-11-24 08:49:17.03992014 +0000 UTC m=+2030.915065313" watchObservedRunningTime="2025-11-24 08:49:17.04238593 +0000 UTC m=+2030.917531073" Nov 24 08:49:26 crc kubenswrapper[4831]: I1124 08:49:26.760299 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:26 crc kubenswrapper[4831]: I1124 08:49:26.761031 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:27 crc kubenswrapper[4831]: I1124 08:49:27.816915 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zvbpc" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" probeResult="failure" output=< Nov 24 08:49:27 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:49:27 crc kubenswrapper[4831]: > Nov 24 08:49:28 crc kubenswrapper[4831]: I1124 08:49:28.401047 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:49:28 crc kubenswrapper[4831]: I1124 08:49:28.401143 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:49:28 crc kubenswrapper[4831]: I1124 08:49:28.401633 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:49:28 crc kubenswrapper[4831]: I1124 08:49:28.402251 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:49:28 crc kubenswrapper[4831]: I1124 08:49:28.402388 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce" gracePeriod=600 Nov 24 08:49:28 crc kubenswrapper[4831]: E1124 08:49:28.589808 4831 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b857b1a_bb8a_4608_8643_e0a94ed82a1b.slice/crio-1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b857b1a_bb8a_4608_8643_e0a94ed82a1b.slice/crio-conmon-1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce.scope\": RecentStats: unable to find data in memory cache]" Nov 24 08:49:29 crc kubenswrapper[4831]: I1124 08:49:29.121596 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce" exitCode=0 Nov 24 08:49:29 crc kubenswrapper[4831]: I1124 08:49:29.121647 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce"} Nov 24 08:49:29 crc kubenswrapper[4831]: I1124 08:49:29.121680 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062"} Nov 24 08:49:29 crc kubenswrapper[4831]: I1124 08:49:29.121702 4831 scope.go:117] "RemoveContainer" containerID="6dbb2252f731ba50044d919e5735a68d22328697945a8b015bd4d63a9ffde2d8" Nov 24 08:49:37 crc kubenswrapper[4831]: I1124 08:49:37.800183 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zvbpc" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" probeResult="failure" output=< Nov 24 08:49:37 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 08:49:37 crc kubenswrapper[4831]: > Nov 24 08:49:46 crc kubenswrapper[4831]: I1124 08:49:46.816917 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:46 crc kubenswrapper[4831]: I1124 08:49:46.868999 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:47 crc kubenswrapper[4831]: I1124 08:49:47.063678 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.288925 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zvbpc" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" containerID="cri-o://6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224" gracePeriod=2 Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.697265 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.859343 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content\") pod \"93868c25-c518-4966-ad6e-9b3ca1d26798\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.859536 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities\") pod \"93868c25-c518-4966-ad6e-9b3ca1d26798\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.859601 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zslmx\" (UniqueName: \"kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx\") pod \"93868c25-c518-4966-ad6e-9b3ca1d26798\" (UID: \"93868c25-c518-4966-ad6e-9b3ca1d26798\") " Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.862408 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities" (OuterVolumeSpecName: "utilities") pod "93868c25-c518-4966-ad6e-9b3ca1d26798" (UID: "93868c25-c518-4966-ad6e-9b3ca1d26798"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.871713 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx" (OuterVolumeSpecName: "kube-api-access-zslmx") pod "93868c25-c518-4966-ad6e-9b3ca1d26798" (UID: "93868c25-c518-4966-ad6e-9b3ca1d26798"). InnerVolumeSpecName "kube-api-access-zslmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.962172 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.962208 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zslmx\" (UniqueName: \"kubernetes.io/projected/93868c25-c518-4966-ad6e-9b3ca1d26798-kube-api-access-zslmx\") on node \"crc\" DevicePath \"\"" Nov 24 08:49:48 crc kubenswrapper[4831]: I1124 08:49:48.978961 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93868c25-c518-4966-ad6e-9b3ca1d26798" (UID: "93868c25-c518-4966-ad6e-9b3ca1d26798"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.064663 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93868c25-c518-4966-ad6e-9b3ca1d26798-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.300871 4831 generic.go:334] "Generic (PLEG): container finished" podID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerID="6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224" exitCode=0 Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.301149 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerDied","Data":"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224"} Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.301225 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zvbpc" event={"ID":"93868c25-c518-4966-ad6e-9b3ca1d26798","Type":"ContainerDied","Data":"1c29c523461afac7bcc9a0c9123312ca5e0467e3b47e53f260c8467454025fde"} Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.301262 4831 scope.go:117] "RemoveContainer" containerID="6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.301283 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zvbpc" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.329223 4831 scope.go:117] "RemoveContainer" containerID="a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.338885 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.350294 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zvbpc"] Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.366596 4831 scope.go:117] "RemoveContainer" containerID="af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.394932 4831 scope.go:117] "RemoveContainer" containerID="6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224" Nov 24 08:49:49 crc kubenswrapper[4831]: E1124 08:49:49.395521 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224\": container with ID starting with 6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224 not found: ID does not exist" containerID="6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.395557 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224"} err="failed to get container status \"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224\": rpc error: code = NotFound desc = could not find container \"6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224\": container with ID starting with 6bb310ef2a25765f850b5cd1aa39c875fa35f849858a8432cc3f44794c262224 not found: ID does not exist" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.395603 4831 scope.go:117] "RemoveContainer" containerID="a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae" Nov 24 08:49:49 crc kubenswrapper[4831]: E1124 08:49:49.396015 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae\": container with ID starting with a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae not found: ID does not exist" containerID="a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.396056 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae"} err="failed to get container status \"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae\": rpc error: code = NotFound desc = could not find container \"a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae\": container with ID starting with a3264e395e4eeb032a3143802685cd607026dc74d2ea4e4378ec7fd52a1e2fae not found: ID does not exist" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.396071 4831 scope.go:117] "RemoveContainer" containerID="af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9" Nov 24 08:49:49 crc kubenswrapper[4831]: E1124 08:49:49.396441 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9\": container with ID starting with af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9 not found: ID does not exist" containerID="af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9" Nov 24 08:49:49 crc kubenswrapper[4831]: I1124 08:49:49.396461 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9"} err="failed to get container status \"af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9\": rpc error: code = NotFound desc = could not find container \"af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9\": container with ID starting with af41d1c6e5629c3d43bd6799c6cd32d2add7766ffd32eb64a9be266082156cd9 not found: ID does not exist" Nov 24 08:49:50 crc kubenswrapper[4831]: I1124 08:49:50.907855 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" path="/var/lib/kubelet/pods/93868c25-c518-4966-ad6e-9b3ca1d26798/volumes" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.241041 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:01 crc kubenswrapper[4831]: E1124 08:51:01.242913 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="extract-utilities" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.243012 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="extract-utilities" Nov 24 08:51:01 crc kubenswrapper[4831]: E1124 08:51:01.243097 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.243171 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" Nov 24 08:51:01 crc kubenswrapper[4831]: E1124 08:51:01.243249 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="extract-content" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.243354 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="extract-content" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.243646 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="93868c25-c518-4966-ad6e-9b3ca1d26798" containerName="registry-server" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.245351 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.256470 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.377285 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjshd\" (UniqueName: \"kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.377918 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.378235 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.480014 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.480346 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjshd\" (UniqueName: \"kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.480506 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.480558 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.480758 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.505677 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjshd\" (UniqueName: \"kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd\") pod \"certified-operators-7n9mq\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.576652 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:01 crc kubenswrapper[4831]: I1124 08:51:01.936597 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:02 crc kubenswrapper[4831]: I1124 08:51:02.653294 4831 generic.go:334] "Generic (PLEG): container finished" podID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerID="2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31" exitCode=0 Nov 24 08:51:02 crc kubenswrapper[4831]: I1124 08:51:02.653399 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerDied","Data":"2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31"} Nov 24 08:51:02 crc kubenswrapper[4831]: I1124 08:51:02.653609 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerStarted","Data":"282cd55dcee65efa36e659193d4dc974e8bca6f26744f7bb2c87849665bdae1e"} Nov 24 08:51:03 crc kubenswrapper[4831]: I1124 08:51:03.663206 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerStarted","Data":"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f"} Nov 24 08:51:04 crc kubenswrapper[4831]: I1124 08:51:04.673979 4831 generic.go:334] "Generic (PLEG): container finished" podID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerID="b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f" exitCode=0 Nov 24 08:51:04 crc kubenswrapper[4831]: I1124 08:51:04.674272 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerDied","Data":"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f"} Nov 24 08:51:05 crc kubenswrapper[4831]: I1124 08:51:05.685690 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerStarted","Data":"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24"} Nov 24 08:51:05 crc kubenswrapper[4831]: I1124 08:51:05.705456 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7n9mq" podStartSLOduration=2.257966766 podStartE2EDuration="4.705438839s" podCreationTimestamp="2025-11-24 08:51:01 +0000 UTC" firstStartedPulling="2025-11-24 08:51:02.655639951 +0000 UTC m=+2136.530785104" lastFinishedPulling="2025-11-24 08:51:05.103112024 +0000 UTC m=+2138.978257177" observedRunningTime="2025-11-24 08:51:05.702099835 +0000 UTC m=+2139.577244988" watchObservedRunningTime="2025-11-24 08:51:05.705438839 +0000 UTC m=+2139.580583992" Nov 24 08:51:11 crc kubenswrapper[4831]: I1124 08:51:11.577460 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:11 crc kubenswrapper[4831]: I1124 08:51:11.578058 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:11 crc kubenswrapper[4831]: I1124 08:51:11.623003 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:11 crc kubenswrapper[4831]: I1124 08:51:11.772116 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:11 crc kubenswrapper[4831]: I1124 08:51:11.851584 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:13 crc kubenswrapper[4831]: I1124 08:51:13.742916 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7n9mq" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="registry-server" containerID="cri-o://492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24" gracePeriod=2 Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.707851 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.752185 4831 generic.go:334] "Generic (PLEG): container finished" podID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerID="492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24" exitCode=0 Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.752227 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerDied","Data":"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24"} Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.752255 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7n9mq" event={"ID":"d2422518-e7fd-4c25-bacf-5b00bda7263e","Type":"ContainerDied","Data":"282cd55dcee65efa36e659193d4dc974e8bca6f26744f7bb2c87849665bdae1e"} Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.752275 4831 scope.go:117] "RemoveContainer" containerID="492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.752420 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7n9mq" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.775774 4831 scope.go:117] "RemoveContainer" containerID="b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.793016 4831 scope.go:117] "RemoveContainer" containerID="2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.816311 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities\") pod \"d2422518-e7fd-4c25-bacf-5b00bda7263e\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.816824 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjshd\" (UniqueName: \"kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd\") pod \"d2422518-e7fd-4c25-bacf-5b00bda7263e\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.816889 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content\") pod \"d2422518-e7fd-4c25-bacf-5b00bda7263e\" (UID: \"d2422518-e7fd-4c25-bacf-5b00bda7263e\") " Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.817167 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities" (OuterVolumeSpecName: "utilities") pod "d2422518-e7fd-4c25-bacf-5b00bda7263e" (UID: "d2422518-e7fd-4c25-bacf-5b00bda7263e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.817271 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.824899 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd" (OuterVolumeSpecName: "kube-api-access-gjshd") pod "d2422518-e7fd-4c25-bacf-5b00bda7263e" (UID: "d2422518-e7fd-4c25-bacf-5b00bda7263e"). InnerVolumeSpecName "kube-api-access-gjshd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.868432 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2422518-e7fd-4c25-bacf-5b00bda7263e" (UID: "d2422518-e7fd-4c25-bacf-5b00bda7263e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.901633 4831 scope.go:117] "RemoveContainer" containerID="492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24" Nov 24 08:51:14 crc kubenswrapper[4831]: E1124 08:51:14.902165 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24\": container with ID starting with 492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24 not found: ID does not exist" containerID="492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.902190 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24"} err="failed to get container status \"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24\": rpc error: code = NotFound desc = could not find container \"492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24\": container with ID starting with 492ddd76869c40489ee7dcf1ab9a935ade1e36f4609fa27c3209d00af617fb24 not found: ID does not exist" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.902209 4831 scope.go:117] "RemoveContainer" containerID="b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f" Nov 24 08:51:14 crc kubenswrapper[4831]: E1124 08:51:14.903989 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f\": container with ID starting with b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f not found: ID does not exist" containerID="b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.904010 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f"} err="failed to get container status \"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f\": rpc error: code = NotFound desc = could not find container \"b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f\": container with ID starting with b7dc1bf2d396b327c310b7476cbc4a7c4577bf4b03b5c3667de9e38a66af0f7f not found: ID does not exist" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.904024 4831 scope.go:117] "RemoveContainer" containerID="2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31" Nov 24 08:51:14 crc kubenswrapper[4831]: E1124 08:51:14.904241 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31\": container with ID starting with 2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31 not found: ID does not exist" containerID="2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.904264 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31"} err="failed to get container status \"2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31\": rpc error: code = NotFound desc = could not find container \"2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31\": container with ID starting with 2dbe13b591bb7a1de156a73d13b6bbb2e09d31c9225df16c7bb2786507601e31 not found: ID does not exist" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.919671 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjshd\" (UniqueName: \"kubernetes.io/projected/d2422518-e7fd-4c25-bacf-5b00bda7263e-kube-api-access-gjshd\") on node \"crc\" DevicePath \"\"" Nov 24 08:51:14 crc kubenswrapper[4831]: I1124 08:51:14.919720 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2422518-e7fd-4c25-bacf-5b00bda7263e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:51:15 crc kubenswrapper[4831]: I1124 08:51:15.076799 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:15 crc kubenswrapper[4831]: I1124 08:51:15.084837 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7n9mq"] Nov 24 08:51:16 crc kubenswrapper[4831]: I1124 08:51:16.906077 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" path="/var/lib/kubelet/pods/d2422518-e7fd-4c25-bacf-5b00bda7263e/volumes" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.324714 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-889c2/must-gather-5hwdk"] Nov 24 08:51:17 crc kubenswrapper[4831]: E1124 08:51:17.325160 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="extract-content" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.325180 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="extract-content" Nov 24 08:51:17 crc kubenswrapper[4831]: E1124 08:51:17.325222 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="extract-utilities" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.325231 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="extract-utilities" Nov 24 08:51:17 crc kubenswrapper[4831]: E1124 08:51:17.325249 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="registry-server" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.325256 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="registry-server" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.325482 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2422518-e7fd-4c25-bacf-5b00bda7263e" containerName="registry-server" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.328416 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.340089 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-889c2"/"openshift-service-ca.crt" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.340129 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-889c2"/"kube-root-ca.crt" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.340403 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-889c2"/"default-dockercfg-5cp54" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.370143 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-889c2/must-gather-5hwdk"] Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.466287 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcqwg\" (UniqueName: \"kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.466580 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.568441 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcqwg\" (UniqueName: \"kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.568879 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.569364 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.585576 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcqwg\" (UniqueName: \"kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg\") pod \"must-gather-5hwdk\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:17 crc kubenswrapper[4831]: I1124 08:51:17.669179 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:51:18 crc kubenswrapper[4831]: I1124 08:51:18.172472 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-889c2/must-gather-5hwdk"] Nov 24 08:51:18 crc kubenswrapper[4831]: W1124 08:51:18.179424 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod091d05b8_2f36_4fb0_bb7c_5258781ddd40.slice/crio-755a4d5c7aafc5bb77bde00b70565ac4f498799c56486a2c8a4abb1e59b63580 WatchSource:0}: Error finding container 755a4d5c7aafc5bb77bde00b70565ac4f498799c56486a2c8a4abb1e59b63580: Status 404 returned error can't find the container with id 755a4d5c7aafc5bb77bde00b70565ac4f498799c56486a2c8a4abb1e59b63580 Nov 24 08:51:18 crc kubenswrapper[4831]: I1124 08:51:18.814685 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/must-gather-5hwdk" event={"ID":"091d05b8-2f36-4fb0-bb7c-5258781ddd40","Type":"ContainerStarted","Data":"755a4d5c7aafc5bb77bde00b70565ac4f498799c56486a2c8a4abb1e59b63580"} Nov 24 08:51:27 crc kubenswrapper[4831]: I1124 08:51:27.903679 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/must-gather-5hwdk" event={"ID":"091d05b8-2f36-4fb0-bb7c-5258781ddd40","Type":"ContainerStarted","Data":"1d55fcb7ea7aa4710ad863397d47d6225b536a90b9cb0218e81f1bb9b9cf7d9b"} Nov 24 08:51:27 crc kubenswrapper[4831]: I1124 08:51:27.904216 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/must-gather-5hwdk" event={"ID":"091d05b8-2f36-4fb0-bb7c-5258781ddd40","Type":"ContainerStarted","Data":"5103476454abde6b6b508466c1199347d39dbab28e9d86eed7d51151bc430175"} Nov 24 08:51:27 crc kubenswrapper[4831]: I1124 08:51:27.920493 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-889c2/must-gather-5hwdk" podStartSLOduration=2.034974071 podStartE2EDuration="10.920455384s" podCreationTimestamp="2025-11-24 08:51:17 +0000 UTC" firstStartedPulling="2025-11-24 08:51:18.182040579 +0000 UTC m=+2152.057185732" lastFinishedPulling="2025-11-24 08:51:27.067521892 +0000 UTC m=+2160.942667045" observedRunningTime="2025-11-24 08:51:27.917278924 +0000 UTC m=+2161.792424097" watchObservedRunningTime="2025-11-24 08:51:27.920455384 +0000 UTC m=+2161.795600557" Nov 24 08:51:28 crc kubenswrapper[4831]: I1124 08:51:28.401552 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:51:28 crc kubenswrapper[4831]: I1124 08:51:28.401601 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:51:30 crc kubenswrapper[4831]: I1124 08:51:30.975732 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-889c2/crc-debug-2qwlw"] Nov 24 08:51:30 crc kubenswrapper[4831]: I1124 08:51:30.977451 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.065939 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.066080 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvcp\" (UniqueName: \"kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.167888 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvcp\" (UniqueName: \"kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.168065 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.168209 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.188000 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvcp\" (UniqueName: \"kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp\") pod \"crc-debug-2qwlw\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.295015 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:51:31 crc kubenswrapper[4831]: I1124 08:51:31.936783 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/crc-debug-2qwlw" event={"ID":"b78783dd-dfb2-42f4-b420-44921b934023","Type":"ContainerStarted","Data":"7e6dc8e33de40b92ac619afbeb0099b15e7438c57d3e0c150e170f6279f99317"} Nov 24 08:51:45 crc kubenswrapper[4831]: I1124 08:51:45.048540 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/crc-debug-2qwlw" event={"ID":"b78783dd-dfb2-42f4-b420-44921b934023","Type":"ContainerStarted","Data":"b7ef2aa2c673120763d54a52a5fd5d714e0474d527cde4bbfdc6fb150a4a77e1"} Nov 24 08:51:45 crc kubenswrapper[4831]: I1124 08:51:45.067227 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-889c2/crc-debug-2qwlw" podStartSLOduration=1.928315582 podStartE2EDuration="15.067205964s" podCreationTimestamp="2025-11-24 08:51:30 +0000 UTC" firstStartedPulling="2025-11-24 08:51:31.330258487 +0000 UTC m=+2165.205403640" lastFinishedPulling="2025-11-24 08:51:44.469148869 +0000 UTC m=+2178.344294022" observedRunningTime="2025-11-24 08:51:45.062604144 +0000 UTC m=+2178.937749297" watchObservedRunningTime="2025-11-24 08:51:45.067205964 +0000 UTC m=+2178.942351117" Nov 24 08:51:58 crc kubenswrapper[4831]: I1124 08:51:58.401414 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:51:58 crc kubenswrapper[4831]: I1124 08:51:58.402055 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:52:11 crc kubenswrapper[4831]: I1124 08:52:11.287377 4831 generic.go:334] "Generic (PLEG): container finished" podID="b78783dd-dfb2-42f4-b420-44921b934023" containerID="b7ef2aa2c673120763d54a52a5fd5d714e0474d527cde4bbfdc6fb150a4a77e1" exitCode=0 Nov 24 08:52:11 crc kubenswrapper[4831]: I1124 08:52:11.287475 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/crc-debug-2qwlw" event={"ID":"b78783dd-dfb2-42f4-b420-44921b934023","Type":"ContainerDied","Data":"b7ef2aa2c673120763d54a52a5fd5d714e0474d527cde4bbfdc6fb150a4a77e1"} Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.413372 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.422379 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host\") pod \"b78783dd-dfb2-42f4-b420-44921b934023\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.422594 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbvcp\" (UniqueName: \"kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp\") pod \"b78783dd-dfb2-42f4-b420-44921b934023\" (UID: \"b78783dd-dfb2-42f4-b420-44921b934023\") " Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.423938 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host" (OuterVolumeSpecName: "host") pod "b78783dd-dfb2-42f4-b420-44921b934023" (UID: "b78783dd-dfb2-42f4-b420-44921b934023"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.438534 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp" (OuterVolumeSpecName: "kube-api-access-hbvcp") pod "b78783dd-dfb2-42f4-b420-44921b934023" (UID: "b78783dd-dfb2-42f4-b420-44921b934023"). InnerVolumeSpecName "kube-api-access-hbvcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.459633 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-889c2/crc-debug-2qwlw"] Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.474608 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-889c2/crc-debug-2qwlw"] Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.526036 4831 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b78783dd-dfb2-42f4-b420-44921b934023-host\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.526368 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbvcp\" (UniqueName: \"kubernetes.io/projected/b78783dd-dfb2-42f4-b420-44921b934023-kube-api-access-hbvcp\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:12 crc kubenswrapper[4831]: I1124 08:52:12.905298 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78783dd-dfb2-42f4-b420-44921b934023" path="/var/lib/kubelet/pods/b78783dd-dfb2-42f4-b420-44921b934023/volumes" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.304394 4831 scope.go:117] "RemoveContainer" containerID="b7ef2aa2c673120763d54a52a5fd5d714e0474d527cde4bbfdc6fb150a4a77e1" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.304475 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-2qwlw" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.984866 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-889c2/crc-debug-77mnm"] Nov 24 08:52:13 crc kubenswrapper[4831]: E1124 08:52:13.985498 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78783dd-dfb2-42f4-b420-44921b934023" containerName="container-00" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.985510 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78783dd-dfb2-42f4-b420-44921b934023" containerName="container-00" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.985699 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78783dd-dfb2-42f4-b420-44921b934023" containerName="container-00" Nov 24 08:52:13 crc kubenswrapper[4831]: I1124 08:52:13.986256 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.054658 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.054765 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljcg9\" (UniqueName: \"kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.156193 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljcg9\" (UniqueName: \"kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.156309 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.156447 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.176840 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljcg9\" (UniqueName: \"kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9\") pod \"crc-debug-77mnm\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:14 crc kubenswrapper[4831]: I1124 08:52:14.301309 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:15 crc kubenswrapper[4831]: I1124 08:52:15.323411 4831 generic.go:334] "Generic (PLEG): container finished" podID="1fa359e7-4780-4059-bbc4-f03ad2c81d13" containerID="4392cc15955abf8edb3fd443b024ac71a1e585aa0f18ca91432571f6a29dec72" exitCode=1 Nov 24 08:52:15 crc kubenswrapper[4831]: I1124 08:52:15.323460 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/crc-debug-77mnm" event={"ID":"1fa359e7-4780-4059-bbc4-f03ad2c81d13","Type":"ContainerDied","Data":"4392cc15955abf8edb3fd443b024ac71a1e585aa0f18ca91432571f6a29dec72"} Nov 24 08:52:15 crc kubenswrapper[4831]: I1124 08:52:15.323740 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/crc-debug-77mnm" event={"ID":"1fa359e7-4780-4059-bbc4-f03ad2c81d13","Type":"ContainerStarted","Data":"df5138540b962cc1ea3f6f0fb595bf9e62cb24b44240fb9daa5920b6e987c228"} Nov 24 08:52:15 crc kubenswrapper[4831]: I1124 08:52:15.384484 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-889c2/crc-debug-77mnm"] Nov 24 08:52:15 crc kubenswrapper[4831]: I1124 08:52:15.393318 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-889c2/crc-debug-77mnm"] Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.456019 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.496300 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljcg9\" (UniqueName: \"kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9\") pod \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.496500 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host\") pod \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\" (UID: \"1fa359e7-4780-4059-bbc4-f03ad2c81d13\") " Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.496966 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host" (OuterVolumeSpecName: "host") pod "1fa359e7-4780-4059-bbc4-f03ad2c81d13" (UID: "1fa359e7-4780-4059-bbc4-f03ad2c81d13"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.513687 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9" (OuterVolumeSpecName: "kube-api-access-ljcg9") pod "1fa359e7-4780-4059-bbc4-f03ad2c81d13" (UID: "1fa359e7-4780-4059-bbc4-f03ad2c81d13"). InnerVolumeSpecName "kube-api-access-ljcg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.597548 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljcg9\" (UniqueName: \"kubernetes.io/projected/1fa359e7-4780-4059-bbc4-f03ad2c81d13-kube-api-access-ljcg9\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.597582 4831 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1fa359e7-4780-4059-bbc4-f03ad2c81d13-host\") on node \"crc\" DevicePath \"\"" Nov 24 08:52:16 crc kubenswrapper[4831]: I1124 08:52:16.911206 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa359e7-4780-4059-bbc4-f03ad2c81d13" path="/var/lib/kubelet/pods/1fa359e7-4780-4059-bbc4-f03ad2c81d13/volumes" Nov 24 08:52:17 crc kubenswrapper[4831]: I1124 08:52:17.344428 4831 scope.go:117] "RemoveContainer" containerID="4392cc15955abf8edb3fd443b024ac71a1e585aa0f18ca91432571f6a29dec72" Nov 24 08:52:17 crc kubenswrapper[4831]: I1124 08:52:17.344700 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/crc-debug-77mnm" Nov 24 08:52:28 crc kubenswrapper[4831]: I1124 08:52:28.401495 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:52:28 crc kubenswrapper[4831]: I1124 08:52:28.402086 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 08:52:28 crc kubenswrapper[4831]: I1124 08:52:28.402158 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 08:52:28 crc kubenswrapper[4831]: I1124 08:52:28.402872 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 08:52:28 crc kubenswrapper[4831]: I1124 08:52:28.402917 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" gracePeriod=600 Nov 24 08:52:28 crc kubenswrapper[4831]: E1124 08:52:28.662699 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:52:29 crc kubenswrapper[4831]: I1124 08:52:29.453061 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" exitCode=0 Nov 24 08:52:29 crc kubenswrapper[4831]: I1124 08:52:29.453120 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062"} Nov 24 08:52:29 crc kubenswrapper[4831]: I1124 08:52:29.453151 4831 scope.go:117] "RemoveContainer" containerID="1ac677fe811b246ced9d930a360fde24ab54c2f94308c9889ab78ea39910f2ce" Nov 24 08:52:29 crc kubenswrapper[4831]: I1124 08:52:29.453846 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:52:29 crc kubenswrapper[4831]: E1124 08:52:29.454299 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:52:42 crc kubenswrapper[4831]: I1124 08:52:42.894128 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:52:42 crc kubenswrapper[4831]: E1124 08:52:42.894883 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.334092 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d4cd7dbc6-zgxq2_52ce26fc-b850-4fb1-9539-e9599f080475/barbican-api/0.log" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.550090 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6d4cd7dbc6-zgxq2_52ce26fc-b850-4fb1-9539-e9599f080475/barbican-api-log/0.log" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.736820 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bcf554d66-64cr7_6b8a5828-9ea3-402e-812d-5dbbbe749087/barbican-keystone-listener/0.log" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.831075 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bcf554d66-64cr7_6b8a5828-9ea3-402e-812d-5dbbbe749087/barbican-keystone-listener-log/0.log" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.893922 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:52:57 crc kubenswrapper[4831]: E1124 08:52:57.894168 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:52:57 crc kubenswrapper[4831]: I1124 08:52:57.941386 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64c5f59669-7kdzn_1087935c-a660-42ae-bc8e-dc8c91bbb414/barbican-worker/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.021305 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-64c5f59669-7kdzn_1087935c-a660-42ae-bc8e-dc8c91bbb414/barbican-worker-log/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.340833 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-cb6fl_fcc403cb-362a-408c-a86b-842ee0b5f9f7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.410929 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_21235403-7ca3-4341-9f1e-661e7e833b32/ceilometer-central-agent/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.488447 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_21235403-7ca3-4341-9f1e-661e7e833b32/ceilometer-notification-agent/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.567551 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_21235403-7ca3-4341-9f1e-661e7e833b32/proxy-httpd/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.613963 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_21235403-7ca3-4341-9f1e-661e7e833b32/sg-core/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.755845 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-h2zfq_8200ad7a-4b9e-46f0-861b-1fc331d347e3/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:52:58 crc kubenswrapper[4831]: I1124 08:52:58.894782 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7795091a-65df-4971-b520-e9cdaa870328/cinder-api/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.064936 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_7795091a-65df-4971-b520-e9cdaa870328/cinder-api-log/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.155861 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c13cfbd9-7d6f-4c2d-96aa-5de94bda5352/cinder-scheduler/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.203363 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_c13cfbd9-7d6f-4c2d-96aa-5de94bda5352/probe/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.321069 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-q8s2v_867472cb-8f0a-4f19-88c5-090bb1165a22/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.553868 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-nlvdm_88cf07ba-1cec-473e-8306-9b1594f166f5/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.610469 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-pw8bh_af25e9d0-61ba-4758-8c40-1df95e214067/init/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.938744 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-pw8bh_af25e9d0-61ba-4758-8c40-1df95e214067/dnsmasq-dns/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.948935 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-667ff9c869-pw8bh_af25e9d0-61ba-4758-8c40-1df95e214067/init/0.log" Nov 24 08:52:59 crc kubenswrapper[4831]: I1124 08:52:59.991197 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jl4hn_23122f3c-2de2-4a17-b899-d71ea63b8b7f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:53:00 crc kubenswrapper[4831]: I1124 08:53:00.258567 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-77b5948bf-z797k_84b88081-1ab2-4e86-a37d-3b4687aeb91e/keystone-api/0.log" Nov 24 08:53:00 crc kubenswrapper[4831]: I1124 08:53:00.313122 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_76ead5ac-ef11-4309-b031-10ea621a212e/kube-state-metrics/0.log" Nov 24 08:53:00 crc kubenswrapper[4831]: I1124 08:53:00.696237 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79dd578589-cbqvb_a511397f-1a08-4819-8f25-4435c4e5fe6a/neutron-api/0.log" Nov 24 08:53:00 crc kubenswrapper[4831]: I1124 08:53:00.796935 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79dd578589-cbqvb_a511397f-1a08-4819-8f25-4435c4e5fe6a/neutron-httpd/0.log" Nov 24 08:53:01 crc kubenswrapper[4831]: I1124 08:53:01.169350 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_de07ebb9-282e-4dd6-9a4f-8267816e827d/nova-api-api/0.log" Nov 24 08:53:01 crc kubenswrapper[4831]: I1124 08:53:01.320110 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_de07ebb9-282e-4dd6-9a4f-8267816e827d/nova-api-log/0.log" Nov 24 08:53:01 crc kubenswrapper[4831]: I1124 08:53:01.602881 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d798294b-fc7d-4ec4-8ad6-5bf73c68144d/nova-cell0-conductor-conductor/0.log" Nov 24 08:53:01 crc kubenswrapper[4831]: I1124 08:53:01.863634 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4aba5ba5-56d4-4883-af9e-9278046b3285/nova-cell1-conductor-conductor/0.log" Nov 24 08:53:02 crc kubenswrapper[4831]: I1124 08:53:02.086371 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4112313c-5cfc-4e0a-bf38-cd6cf34f309e/nova-cell1-novncproxy-novncproxy/0.log" Nov 24 08:53:02 crc kubenswrapper[4831]: I1124 08:53:02.438607 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_86f061df-2ac1-4323-b6ce-b6c710d98793/nova-metadata-log/0.log" Nov 24 08:53:02 crc kubenswrapper[4831]: I1124 08:53:02.644574 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6df77d90-f9a7-4653-83e0-d465588ed50e/nova-scheduler-scheduler/0.log" Nov 24 08:53:02 crc kubenswrapper[4831]: I1124 08:53:02.787218 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7d136719-7712-457e-a66c-d4f349c8c341/mysql-bootstrap/0.log" Nov 24 08:53:02 crc kubenswrapper[4831]: I1124 08:53:02.952381 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_86f061df-2ac1-4323-b6ce-b6c710d98793/nova-metadata-metadata/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.078129 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7d136719-7712-457e-a66c-d4f349c8c341/mysql-bootstrap/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.087925 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_7d136719-7712-457e-a66c-d4f349c8c341/galera/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.318553 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2/mysql-bootstrap/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.514950 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2/galera/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.572135 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_e2511d1a-25c1-4825-a4e3-bdabb5dfbcf2/mysql-bootstrap/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.642152 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8db8fec2-ed06-4ad5-8fcf-a491ff18abd1/openstackclient/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.911445 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4wbwz_2c454278-b245-4dba-a3e5-0fbb86db5ab0/ovsdb-server-init/0.log" Nov 24 08:53:03 crc kubenswrapper[4831]: I1124 08:53:03.922783 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-g88jf_23b22b28-ece1-400e-9e58-a34fa7392bfe/openstack-network-exporter/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.272622 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4wbwz_2c454278-b245-4dba-a3e5-0fbb86db5ab0/ovsdb-server/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.298142 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4wbwz_2c454278-b245-4dba-a3e5-0fbb86db5ab0/ovs-vswitchd/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.314281 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-4wbwz_2c454278-b245-4dba-a3e5-0fbb86db5ab0/ovsdb-server-init/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.577917 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-p87rh_71dd0aae-5932-4334-9639-49a91209160c/ovn-controller/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.626737 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_901cd1f1-9eda-4434-abd5-d128f1cd2511/openstack-network-exporter/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.769968 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_901cd1f1-9eda-4434-abd5-d128f1cd2511/ovn-northd/0.log" Nov 24 08:53:04 crc kubenswrapper[4831]: I1124 08:53:04.995991 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9974ab31-9887-4acf-8491-0e8b3aec7026/openstack-network-exporter/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.000291 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9974ab31-9887-4acf-8491-0e8b3aec7026/ovsdbserver-nb/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.160782 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b44afa71-a830-4284-87a1-d42b96db03a9/openstack-network-exporter/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.362916 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_b44afa71-a830-4284-87a1-d42b96db03a9/ovsdbserver-sb/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.469963 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f67674df6-sk4ws_610d7193-b921-486e-8d2c-aa897a589c30/placement-api/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.622685 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6f67674df6-sk4ws_610d7193-b921-486e-8d2c-aa897a589c30/placement-log/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.706933 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ee71bf1-4264-4480-843b-04bef17b232a/setup-container/0.log" Nov 24 08:53:05 crc kubenswrapper[4831]: I1124 08:53:05.951362 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ee71bf1-4264-4480-843b-04bef17b232a/setup-container/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.005239 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d6af69e-ffd4-4001-b938-18df9e23981c/setup-container/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.014736 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ee71bf1-4264-4480-843b-04bef17b232a/rabbitmq/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.174814 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d6af69e-ffd4-4001-b938-18df9e23981c/setup-container/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.357899 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_3d6af69e-ffd4-4001-b938-18df9e23981c/rabbitmq/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.450696 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jcx7d_95fe9895-4e23-4cb9-9eb1-1a76a94a31d7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.614221 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lrsdt_f56f5009-34b0-45e5-b303-1c7b123e7600/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.774450 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zmdx9_aa8c7ab9-41a6-412f-9abf-bc4105b768d9/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:53:06 crc kubenswrapper[4831]: I1124 08:53:06.918834 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-qbqmz_5fe18501-edce-4461-ab81-04b861259785/ssh-known-hosts-edpm-deployment/0.log" Nov 24 08:53:07 crc kubenswrapper[4831]: I1124 08:53:07.105618 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wj8xq_450b41f3-ecf5-4f17-a774-ade769d2fa1f/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 24 08:53:09 crc kubenswrapper[4831]: I1124 08:53:09.058672 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_cfda28f8-5deb-43df-9548-3dabfc21d6c0/memcached/0.log" Nov 24 08:53:11 crc kubenswrapper[4831]: I1124 08:53:11.893737 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:53:11 crc kubenswrapper[4831]: E1124 08:53:11.895374 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:53:22 crc kubenswrapper[4831]: I1124 08:53:22.893934 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:53:22 crc kubenswrapper[4831]: E1124 08:53:22.894704 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.560337 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-x7xzk_78bce3ee-0478-4a09-9568-9e62c29dc754/manager/0.log" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.567808 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-75fb479bcc-x7xzk_78bce3ee-0478-4a09-9568-9e62c29dc754/kube-rbac-proxy/0.log" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.773125 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/util/0.log" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.937596 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/pull/0.log" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.953576 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/util/0.log" Nov 24 08:53:28 crc kubenswrapper[4831]: I1124 08:53:28.999252 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/pull/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.154360 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/util/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.156156 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/pull/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.185337 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bdc0f47e712ce274259eb0b98f32b4757252b1463948c997fbf73fc467q228r_ebab0eb6-34eb-480f-8d73-d30997857d21/extract/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.363565 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-994pv_156abc13-0fab-4e2c-a786-98ddf7c340b5/kube-rbac-proxy/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.401421 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6498cbf48f-994pv_156abc13-0fab-4e2c-a786-98ddf7c340b5/manager/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.512195 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-bdfgj_cdc653a8-5c91-4da4-8489-380f2fd0b05d/kube-rbac-proxy/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.595620 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-767ccfd65f-bdfgj_cdc653a8-5c91-4da4-8489-380f2fd0b05d/manager/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.707879 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-xcdjn_ed86dbad-6726-46f7-a11f-feb0647ab7c5/kube-rbac-proxy/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.825359 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7969689c84-xcdjn_ed86dbad-6726-46f7-a11f-feb0647ab7c5/manager/0.log" Nov 24 08:53:29 crc kubenswrapper[4831]: I1124 08:53:29.945199 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-ppdhf_af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc/kube-rbac-proxy/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.047117 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-56f54d6746-ppdhf_af4f58a0-ae5b-4795-a0dc-0ced9e0b9efc/manager/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.092166 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-rshdg_e94a7743-730d-4259-a584-59d1034cd654/kube-rbac-proxy/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.172205 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-598f69df5d-rshdg_e94a7743-730d-4259-a584-59d1034cd654/manager/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.305809 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-m5tlh_054589e5-1de8-4a64-b5ad-2d1688cf8219/kube-rbac-proxy/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.505521 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6dd8864d7c-m5tlh_054589e5-1de8-4a64-b5ad-2d1688cf8219/manager/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.566270 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-rrtwx_4d9702a6-a116-4598-bc95-14c7023cacb0/manager/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.575161 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-99b499f4-rrtwx_4d9702a6-a116-4598-bc95-14c7023cacb0/kube-rbac-proxy/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.735916 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-zz4ns_8bab5961-2a79-482a-8935-b3f84b6e5fbf/kube-rbac-proxy/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.847955 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7454b96578-zz4ns_8bab5961-2a79-482a-8935-b3f84b6e5fbf/manager/0.log" Nov 24 08:53:30 crc kubenswrapper[4831]: I1124 08:53:30.960042 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-zftw6_021fa54a-79c0-483a-96af-9f154175eb2f/kube-rbac-proxy/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.063173 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58f887965d-zftw6_021fa54a-79c0-483a-96af-9f154175eb2f/manager/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.128448 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-zp88s_37a35a63-0850-4615-afe3-905c1218bbcd/kube-rbac-proxy/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.222926 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-54b5986bb8-zp88s_37a35a63-0850-4615-afe3-905c1218bbcd/manager/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.358859 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-hv6f6_e14d1d45-f2ee-4f9d-8260-e10bf3c8333c/kube-rbac-proxy/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.390766 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78bd47f458-hv6f6_e14d1d45-f2ee-4f9d-8260-e10bf3c8333c/manager/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.563430 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-n45zz_4bc8600a-62b5-4733-90cc-b08b66b605fa/kube-rbac-proxy/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.649533 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-cfbb9c588-n45zz_4bc8600a-62b5-4733-90cc-b08b66b605fa/manager/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.766861 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-n8nkt_14cc197a-4c42-46dc-89c6-941dd1997aa8/kube-rbac-proxy/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.844422 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-54cfbf4c7d-n8nkt_14cc197a-4c42-46dc-89c6-941dd1997aa8/manager/0.log" Nov 24 08:53:31 crc kubenswrapper[4831]: I1124 08:53:31.920412 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n_7531434a-d014-4bcb-9765-dd8ca784de62/kube-rbac-proxy/0.log" Nov 24 08:53:32 crc kubenswrapper[4831]: I1124 08:53:32.002708 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-8c7444f48-w7j2n_7531434a-d014-4bcb-9765-dd8ca784de62/manager/0.log" Nov 24 08:53:32 crc kubenswrapper[4831]: I1124 08:53:32.115897 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5dbdc8d64b-pcjw6_145ad046-da71-432f-9b5d-0f2c599c85e0/kube-rbac-proxy/0.log" Nov 24 08:53:32 crc kubenswrapper[4831]: I1124 08:53:32.425733 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84f6799855-frnsm_2940bb03-68aa-433e-a3c2-290fd7993dcd/kube-rbac-proxy/0.log" Nov 24 08:53:32 crc kubenswrapper[4831]: I1124 08:53:32.629264 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84f6799855-frnsm_2940bb03-68aa-433e-a3c2-290fd7993dcd/operator/0.log" Nov 24 08:53:32 crc kubenswrapper[4831]: I1124 08:53:32.877449 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xqbxv_f52364b5-07f0-41a4-a6fb-fa51f8ebd0ec/registry-server/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.072413 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-4c6fw_907add29-1677-495e-b968-432dd37c5e4e/kube-rbac-proxy/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.090146 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5dbdc8d64b-pcjw6_145ad046-da71-432f-9b5d-0f2c599c85e0/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.114097 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-54fc5f65b7-4c6fw_907add29-1677-495e-b968-432dd37c5e4e/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.211925 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-md74d_e7c48858-2602-4763-b69e-96a4f7db47f5/kube-rbac-proxy/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.326825 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b797b8dff-md74d_e7c48858-2602-4763-b69e-96a4f7db47f5/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.371950 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-52vxr_82cf04aa-e08a-4626-ba4c-8c2321aaeb13/operator/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.509160 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-85c8r_905703f5-8abe-45db-8e0e-a01369ef3413/kube-rbac-proxy/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.620406 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-d656998f4-85c8r_905703f5-8abe-45db-8e0e-a01369ef3413/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.710616 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4s2q2_863fe05e-d2e1-47be-b0ab-1d69c2455dc1/kube-rbac-proxy/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.793766 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6d4bf84b58-4s2q2_863fe05e-d2e1-47be-b0ab-1d69c2455dc1/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.880354 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-9d55q_0ee539a7-3c8d-47b9-bf94-bcecc7ff9333/kube-rbac-proxy/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.893267 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:53:33 crc kubenswrapper[4831]: E1124 08:53:33.893511 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.947773 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-b4c496f69-9d55q_0ee539a7-3c8d-47b9-bf94-bcecc7ff9333/manager/0.log" Nov 24 08:53:33 crc kubenswrapper[4831]: I1124 08:53:33.982980 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-k5tgn_03d64575-b4f3-4b39-9fb8-d9597cea126c/kube-rbac-proxy/0.log" Nov 24 08:53:34 crc kubenswrapper[4831]: I1124 08:53:34.052257 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-8c6448b9f-k5tgn_03d64575-b4f3-4b39-9fb8-d9597cea126c/manager/0.log" Nov 24 08:53:47 crc kubenswrapper[4831]: I1124 08:53:47.893682 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:53:47 crc kubenswrapper[4831]: E1124 08:53:47.894646 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:53:50 crc kubenswrapper[4831]: I1124 08:53:50.324159 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rxwsw_ce29f16d-07cb-4d33-95bd-035d6d012ad2/control-plane-machine-set-operator/0.log" Nov 24 08:53:50 crc kubenswrapper[4831]: I1124 08:53:50.540444 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qkrmn_abe7bf71-f32f-4394-93cc-4e3157327c5a/machine-api-operator/0.log" Nov 24 08:53:50 crc kubenswrapper[4831]: I1124 08:53:50.546246 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-qkrmn_abe7bf71-f32f-4394-93cc-4e3157327c5a/kube-rbac-proxy/0.log" Nov 24 08:54:02 crc kubenswrapper[4831]: I1124 08:54:02.847741 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-clth9_8d75df94-e74a-48b4-b04b-b466f484257d/cert-manager-controller/0.log" Nov 24 08:54:02 crc kubenswrapper[4831]: I1124 08:54:02.893712 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:54:02 crc kubenswrapper[4831]: E1124 08:54:02.894005 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:54:03 crc kubenswrapper[4831]: I1124 08:54:03.048346 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ppzf6_c56371b9-cc42-4189-9295-35e64d578865/cert-manager-cainjector/0.log" Nov 24 08:54:03 crc kubenswrapper[4831]: I1124 08:54:03.137998 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-j4gh9_65536f0e-02d6-4cc9-bb72-31cfcc963cdb/cert-manager-webhook/0.log" Nov 24 08:54:14 crc kubenswrapper[4831]: I1124 08:54:14.893771 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:54:14 crc kubenswrapper[4831]: E1124 08:54:14.894527 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.118304 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-cq8tq_5271f267-b380-4526-b535-444b888d87ec/nmstate-console-plugin/0.log" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.366655 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-jv2tr_11078905-8b2b-4454-890d-11a2f640e692/nmstate-handler/0.log" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.423851 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-qpht7_76c1bc31-0ee2-44c8-a2cf-98d66951075c/kube-rbac-proxy/0.log" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.436354 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-qpht7_76c1bc31-0ee2-44c8-a2cf-98d66951075c/nmstate-metrics/0.log" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.623748 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-h7x4c_641a12f7-d840-452b-9d69-4a38b4bb63d7/nmstate-operator/0.log" Nov 24 08:54:15 crc kubenswrapper[4831]: I1124 08:54:15.727532 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-9wd4j_309f3ff6-5b90-44e6-97d5-4e3fbb794b75/nmstate-webhook/0.log" Nov 24 08:54:29 crc kubenswrapper[4831]: I1124 08:54:29.784791 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9d24d_1b0fec8b-1645-468f-acf9-4fdfff8b701b/kube-rbac-proxy/0.log" Nov 24 08:54:29 crc kubenswrapper[4831]: I1124 08:54:29.893524 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:54:29 crc kubenswrapper[4831]: E1124 08:54:29.893901 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:54:29 crc kubenswrapper[4831]: I1124 08:54:29.957667 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-frr-files/0.log" Nov 24 08:54:29 crc kubenswrapper[4831]: I1124 08:54:29.983875 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-9d24d_1b0fec8b-1645-468f-acf9-4fdfff8b701b/controller/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.202300 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-reloader/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.257269 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-frr-files/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.275027 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-metrics/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.322773 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-reloader/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.500744 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-frr-files/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.522474 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-reloader/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.552650 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-metrics/0.log" Nov 24 08:54:30 crc kubenswrapper[4831]: I1124 08:54:30.558682 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-metrics/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.106558 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-frr-files/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.139814 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-reloader/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.147050 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/controller/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.185378 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/cp-metrics/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.383863 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/frr-metrics/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.410497 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/kube-rbac-proxy/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.448460 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/kube-rbac-proxy-frr/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.611954 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/reloader/0.log" Nov 24 08:54:31 crc kubenswrapper[4831]: I1124 08:54:31.820418 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-q4hq4_24e6b9b3-4e9b-4410-bdac-1e359d01375a/frr-k8s-webhook-server/0.log" Nov 24 08:54:32 crc kubenswrapper[4831]: I1124 08:54:32.053839 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7f799f4d7b-m5b2d_7bdd9aa7-6f5e-438e-8646-f05b704b60a6/manager/0.log" Nov 24 08:54:32 crc kubenswrapper[4831]: I1124 08:54:32.214944 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jmst2_6aab2a82-5013-4c8e-8f08-3f24223a5e3e/frr/0.log" Nov 24 08:54:32 crc kubenswrapper[4831]: I1124 08:54:32.251763 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d9bd5646d-74m72_775f3c57-baa8-4e13-8126-eee5194dc2ae/webhook-server/0.log" Nov 24 08:54:32 crc kubenswrapper[4831]: I1124 08:54:32.492784 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-slgk5_03acb26f-311e-456c-a877-dfb4f9439834/kube-rbac-proxy/0.log" Nov 24 08:54:32 crc kubenswrapper[4831]: I1124 08:54:32.830668 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-slgk5_03acb26f-311e-456c-a877-dfb4f9439834/speaker/0.log" Nov 24 08:54:43 crc kubenswrapper[4831]: I1124 08:54:43.895574 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:54:43 crc kubenswrapper[4831]: E1124 08:54:43.896272 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:54:45 crc kubenswrapper[4831]: I1124 08:54:45.847568 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/util/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.000310 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/util/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.070703 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/pull/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.130096 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/pull/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.370427 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/util/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.435364 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/extract/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.493354 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebd9dp_72058868-a5d9-4a81-a6e7-ec9e95140372/pull/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.637229 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-utilities/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.830629 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-utilities/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.882977 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-content/0.log" Nov 24 08:54:46 crc kubenswrapper[4831]: I1124 08:54:46.909596 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-content/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.057956 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-utilities/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.219965 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/extract-content/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.415355 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-utilities/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.468955 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-4jlh4_dc07c012-a5d2-4527-a476-87ff3aae92d8/registry-server/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.601194 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-utilities/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.633464 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-content/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.638772 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-content/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.932700 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-utilities/0.log" Nov 24 08:54:47 crc kubenswrapper[4831]: I1124 08:54:47.999570 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/extract-content/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.261133 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/util/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.288867 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wm7hx_8b8ea8fd-6b03-4105-b88f-0b134bc1bd0f/registry-server/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.515866 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/pull/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.541055 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/pull/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.545233 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/util/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.727006 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/pull/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.763970 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/util/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.807053 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6vlq7m_9b86d122-0df8-4bd9-9a18-21413be5e335/extract/0.log" Nov 24 08:54:48 crc kubenswrapper[4831]: I1124 08:54:48.971699 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-fcz4s_749f8c14-287b-497c-ad5e-cdc9c45b30fe/marketplace-operator/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.064448 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-utilities/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.344420 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-content/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.374999 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-utilities/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.409407 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-content/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.532631 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-utilities/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.580996 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/extract-content/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.746105 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-t6xrr_0e1b1f5e-e348-4846-8089-90c6888f51fa/registry-server/0.log" Nov 24 08:54:49 crc kubenswrapper[4831]: I1124 08:54:49.832907 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-utilities/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.085557 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-utilities/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.088241 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-content/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.102161 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-content/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.323448 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-utilities/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.342410 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/extract-content/0.log" Nov 24 08:54:50 crc kubenswrapper[4831]: I1124 08:54:50.640653 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9d6pp_18b44bbb-ad72-4564-b393-a90ee760fa50/registry-server/0.log" Nov 24 08:54:54 crc kubenswrapper[4831]: I1124 08:54:54.895029 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:54:54 crc kubenswrapper[4831]: E1124 08:54:54.895795 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:55:06 crc kubenswrapper[4831]: I1124 08:55:06.913501 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:55:06 crc kubenswrapper[4831]: E1124 08:55:06.915186 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:55:20 crc kubenswrapper[4831]: I1124 08:55:20.894502 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:55:20 crc kubenswrapper[4831]: E1124 08:55:20.895509 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.297412 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:27 crc kubenswrapper[4831]: E1124 08:55:27.300716 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa359e7-4780-4059-bbc4-f03ad2c81d13" containerName="container-00" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.300747 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa359e7-4780-4059-bbc4-f03ad2c81d13" containerName="container-00" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.300970 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa359e7-4780-4059-bbc4-f03ad2c81d13" containerName="container-00" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.307389 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.310718 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.406591 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf8dn\" (UniqueName: \"kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.406668 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.406773 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.507959 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.508130 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf8dn\" (UniqueName: \"kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.508176 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.510642 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.510902 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.539184 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf8dn\" (UniqueName: \"kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn\") pod \"redhat-marketplace-77ncv\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.640515 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:27 crc kubenswrapper[4831]: I1124 08:55:27.987610 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:28 crc kubenswrapper[4831]: I1124 08:55:28.284022 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerID="47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc" exitCode=0 Nov 24 08:55:28 crc kubenswrapper[4831]: I1124 08:55:28.284223 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerDied","Data":"47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc"} Nov 24 08:55:28 crc kubenswrapper[4831]: I1124 08:55:28.284340 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerStarted","Data":"99df467f0424ae2e4383a2ef81270cbee886d68f6c3ab2d7ac10c9dcf4076bba"} Nov 24 08:55:28 crc kubenswrapper[4831]: I1124 08:55:28.286127 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 08:55:29 crc kubenswrapper[4831]: I1124 08:55:29.294709 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerStarted","Data":"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850"} Nov 24 08:55:30 crc kubenswrapper[4831]: I1124 08:55:30.303291 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerID="ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850" exitCode=0 Nov 24 08:55:30 crc kubenswrapper[4831]: I1124 08:55:30.303392 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerDied","Data":"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850"} Nov 24 08:55:31 crc kubenswrapper[4831]: I1124 08:55:31.314072 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerStarted","Data":"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347"} Nov 24 08:55:31 crc kubenswrapper[4831]: I1124 08:55:31.342759 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-77ncv" podStartSLOduration=1.782278448 podStartE2EDuration="4.342737161s" podCreationTimestamp="2025-11-24 08:55:27 +0000 UTC" firstStartedPulling="2025-11-24 08:55:28.285900432 +0000 UTC m=+2402.161045585" lastFinishedPulling="2025-11-24 08:55:30.846359155 +0000 UTC m=+2404.721504298" observedRunningTime="2025-11-24 08:55:31.336266486 +0000 UTC m=+2405.211411649" watchObservedRunningTime="2025-11-24 08:55:31.342737161 +0000 UTC m=+2405.217882324" Nov 24 08:55:35 crc kubenswrapper[4831]: I1124 08:55:35.894187 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:55:35 crc kubenswrapper[4831]: E1124 08:55:35.894966 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:55:37 crc kubenswrapper[4831]: I1124 08:55:37.640907 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:37 crc kubenswrapper[4831]: I1124 08:55:37.641252 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:37 crc kubenswrapper[4831]: I1124 08:55:37.697677 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:38 crc kubenswrapper[4831]: I1124 08:55:38.430542 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:38 crc kubenswrapper[4831]: I1124 08:55:38.479877 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.394822 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-77ncv" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="registry-server" containerID="cri-o://81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347" gracePeriod=2 Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.858938 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.950849 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf8dn\" (UniqueName: \"kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn\") pod \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.952016 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content\") pod \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.952819 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities" (OuterVolumeSpecName: "utilities") pod "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" (UID: "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.957566 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities\") pod \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\" (UID: \"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db\") " Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.957780 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn" (OuterVolumeSpecName: "kube-api-access-mf8dn") pod "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" (UID: "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db"). InnerVolumeSpecName "kube-api-access-mf8dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.958594 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf8dn\" (UniqueName: \"kubernetes.io/projected/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-kube-api-access-mf8dn\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.958620 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:40 crc kubenswrapper[4831]: I1124 08:55:40.972731 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" (UID: "5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.060708 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.404711 4831 generic.go:334] "Generic (PLEG): container finished" podID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerID="81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347" exitCode=0 Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.404890 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerDied","Data":"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347"} Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.405950 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-77ncv" event={"ID":"5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db","Type":"ContainerDied","Data":"99df467f0424ae2e4383a2ef81270cbee886d68f6c3ab2d7ac10c9dcf4076bba"} Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.406035 4831 scope.go:117] "RemoveContainer" containerID="81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.404964 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-77ncv" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.438635 4831 scope.go:117] "RemoveContainer" containerID="ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.446620 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.457465 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-77ncv"] Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.508697 4831 scope.go:117] "RemoveContainer" containerID="47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.534083 4831 scope.go:117] "RemoveContainer" containerID="81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347" Nov 24 08:55:41 crc kubenswrapper[4831]: E1124 08:55:41.536876 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347\": container with ID starting with 81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347 not found: ID does not exist" containerID="81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.536922 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347"} err="failed to get container status \"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347\": rpc error: code = NotFound desc = could not find container \"81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347\": container with ID starting with 81f6ca33e041c6c83f2c81a5782e5acba7ff12278156b49330ed90b30a931347 not found: ID does not exist" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.536951 4831 scope.go:117] "RemoveContainer" containerID="ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850" Nov 24 08:55:41 crc kubenswrapper[4831]: E1124 08:55:41.537273 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850\": container with ID starting with ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850 not found: ID does not exist" containerID="ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.537305 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850"} err="failed to get container status \"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850\": rpc error: code = NotFound desc = could not find container \"ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850\": container with ID starting with ed8977620e0d768cb4b4e785af3b6b1b54ec12fe2a63324dd280506af4ac7850 not found: ID does not exist" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.537341 4831 scope.go:117] "RemoveContainer" containerID="47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc" Nov 24 08:55:41 crc kubenswrapper[4831]: E1124 08:55:41.537888 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc\": container with ID starting with 47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc not found: ID does not exist" containerID="47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc" Nov 24 08:55:41 crc kubenswrapper[4831]: I1124 08:55:41.537929 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc"} err="failed to get container status \"47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc\": rpc error: code = NotFound desc = could not find container \"47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc\": container with ID starting with 47d743ddc2b40e8009520770817bc33f82728be76789dd0b368ba4bc73f74bcc not found: ID does not exist" Nov 24 08:55:42 crc kubenswrapper[4831]: I1124 08:55:42.907850 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" path="/var/lib/kubelet/pods/5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db/volumes" Nov 24 08:55:48 crc kubenswrapper[4831]: I1124 08:55:48.894003 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:55:48 crc kubenswrapper[4831]: E1124 08:55:48.894815 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:55:59 crc kubenswrapper[4831]: I1124 08:55:59.895461 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:55:59 crc kubenswrapper[4831]: E1124 08:55:59.896293 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:56:12 crc kubenswrapper[4831]: I1124 08:56:12.893202 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:56:12 crc kubenswrapper[4831]: E1124 08:56:12.893969 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:56:25 crc kubenswrapper[4831]: I1124 08:56:25.892836 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:56:25 crc kubenswrapper[4831]: E1124 08:56:25.895632 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:56:38 crc kubenswrapper[4831]: I1124 08:56:38.895349 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:56:38 crc kubenswrapper[4831]: E1124 08:56:38.896374 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:56:49 crc kubenswrapper[4831]: I1124 08:56:49.893830 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:56:49 crc kubenswrapper[4831]: E1124 08:56:49.894860 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:56:53 crc kubenswrapper[4831]: I1124 08:56:53.014879 4831 generic.go:334] "Generic (PLEG): container finished" podID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerID="5103476454abde6b6b508466c1199347d39dbab28e9d86eed7d51151bc430175" exitCode=0 Nov 24 08:56:53 crc kubenswrapper[4831]: I1124 08:56:53.015046 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-889c2/must-gather-5hwdk" event={"ID":"091d05b8-2f36-4fb0-bb7c-5258781ddd40","Type":"ContainerDied","Data":"5103476454abde6b6b508466c1199347d39dbab28e9d86eed7d51151bc430175"} Nov 24 08:56:53 crc kubenswrapper[4831]: I1124 08:56:53.017174 4831 scope.go:117] "RemoveContainer" containerID="5103476454abde6b6b508466c1199347d39dbab28e9d86eed7d51151bc430175" Nov 24 08:56:53 crc kubenswrapper[4831]: I1124 08:56:53.790729 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-889c2_must-gather-5hwdk_091d05b8-2f36-4fb0-bb7c-5258781ddd40/gather/0.log" Nov 24 08:57:02 crc kubenswrapper[4831]: I1124 08:57:02.608105 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-889c2/must-gather-5hwdk"] Nov 24 08:57:02 crc kubenswrapper[4831]: I1124 08:57:02.609247 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-889c2/must-gather-5hwdk" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="copy" containerID="cri-o://1d55fcb7ea7aa4710ad863397d47d6225b536a90b9cb0218e81f1bb9b9cf7d9b" gracePeriod=2 Nov 24 08:57:02 crc kubenswrapper[4831]: I1124 08:57:02.620163 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-889c2/must-gather-5hwdk"] Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.135800 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-889c2_must-gather-5hwdk_091d05b8-2f36-4fb0-bb7c-5258781ddd40/copy/0.log" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.136539 4831 generic.go:334] "Generic (PLEG): container finished" podID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerID="1d55fcb7ea7aa4710ad863397d47d6225b536a90b9cb0218e81f1bb9b9cf7d9b" exitCode=143 Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.766580 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-889c2_must-gather-5hwdk_091d05b8-2f36-4fb0-bb7c-5258781ddd40/copy/0.log" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.768729 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.823853 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output\") pod \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.824402 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcqwg\" (UniqueName: \"kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg\") pod \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\" (UID: \"091d05b8-2f36-4fb0-bb7c-5258781ddd40\") " Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.848870 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg" (OuterVolumeSpecName: "kube-api-access-gcqwg") pod "091d05b8-2f36-4fb0-bb7c-5258781ddd40" (UID: "091d05b8-2f36-4fb0-bb7c-5258781ddd40"). InnerVolumeSpecName "kube-api-access-gcqwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.893980 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:57:03 crc kubenswrapper[4831]: E1124 08:57:03.894536 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.929495 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcqwg\" (UniqueName: \"kubernetes.io/projected/091d05b8-2f36-4fb0-bb7c-5258781ddd40-kube-api-access-gcqwg\") on node \"crc\" DevicePath \"\"" Nov 24 08:57:03 crc kubenswrapper[4831]: I1124 08:57:03.992924 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "091d05b8-2f36-4fb0-bb7c-5258781ddd40" (UID: "091d05b8-2f36-4fb0-bb7c-5258781ddd40"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.031691 4831 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/091d05b8-2f36-4fb0-bb7c-5258781ddd40-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.147270 4831 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-889c2_must-gather-5hwdk_091d05b8-2f36-4fb0-bb7c-5258781ddd40/copy/0.log" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.148086 4831 scope.go:117] "RemoveContainer" containerID="1d55fcb7ea7aa4710ad863397d47d6225b536a90b9cb0218e81f1bb9b9cf7d9b" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.148126 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-889c2/must-gather-5hwdk" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.190827 4831 scope.go:117] "RemoveContainer" containerID="5103476454abde6b6b508466c1199347d39dbab28e9d86eed7d51151bc430175" Nov 24 08:57:04 crc kubenswrapper[4831]: I1124 08:57:04.904409 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" path="/var/lib/kubelet/pods/091d05b8-2f36-4fb0-bb7c-5258781ddd40/volumes" Nov 24 08:57:16 crc kubenswrapper[4831]: I1124 08:57:16.899536 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:57:16 crc kubenswrapper[4831]: E1124 08:57:16.900479 4831 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ftxxx_openshift-machine-config-operator(8b857b1a-bb8a-4608-8643-e0a94ed82a1b)\"" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" Nov 24 08:57:30 crc kubenswrapper[4831]: I1124 08:57:30.894497 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 08:57:31 crc kubenswrapper[4831]: I1124 08:57:31.408713 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"a56572d34a7282b53c43d3feef284935414f2592b77e3306608169fee7e6d305"} Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.316163 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:19 crc kubenswrapper[4831]: E1124 08:58:19.318890 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="extract-content" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.318994 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="extract-content" Nov 24 08:58:19 crc kubenswrapper[4831]: E1124 08:58:19.319092 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="registry-server" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.319153 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="registry-server" Nov 24 08:58:19 crc kubenswrapper[4831]: E1124 08:58:19.319231 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="copy" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.319283 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="copy" Nov 24 08:58:19 crc kubenswrapper[4831]: E1124 08:58:19.319366 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="extract-utilities" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.319424 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="extract-utilities" Nov 24 08:58:19 crc kubenswrapper[4831]: E1124 08:58:19.319493 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="gather" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.319550 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="gather" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.320223 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="gather" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.320302 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="091d05b8-2f36-4fb0-bb7c-5258781ddd40" containerName="copy" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.320394 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b31bdc7-b1a9-4eca-952f-a3e0d21fa8db" containerName="registry-server" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.324799 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.378546 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.452020 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.452290 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.452456 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpdfq\" (UniqueName: \"kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.554424 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.554498 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.554529 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpdfq\" (UniqueName: \"kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.554904 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.555422 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.579082 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpdfq\" (UniqueName: \"kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq\") pod \"community-operators-vprtr\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:19 crc kubenswrapper[4831]: I1124 08:58:19.669713 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:20 crc kubenswrapper[4831]: I1124 08:58:20.238694 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:20 crc kubenswrapper[4831]: I1124 08:58:20.814833 4831 generic.go:334] "Generic (PLEG): container finished" podID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerID="7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd" exitCode=0 Nov 24 08:58:20 crc kubenswrapper[4831]: I1124 08:58:20.814958 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerDied","Data":"7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd"} Nov 24 08:58:20 crc kubenswrapper[4831]: I1124 08:58:20.815151 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerStarted","Data":"7cd9e06edcb92dc2975d929b4b0e7970575212ba6db6870309130969048f1bae"} Nov 24 08:58:21 crc kubenswrapper[4831]: I1124 08:58:21.824036 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerStarted","Data":"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515"} Nov 24 08:58:22 crc kubenswrapper[4831]: I1124 08:58:22.835092 4831 generic.go:334] "Generic (PLEG): container finished" podID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerID="972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515" exitCode=0 Nov 24 08:58:22 crc kubenswrapper[4831]: I1124 08:58:22.835139 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerDied","Data":"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515"} Nov 24 08:58:23 crc kubenswrapper[4831]: I1124 08:58:23.845936 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerStarted","Data":"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581"} Nov 24 08:58:23 crc kubenswrapper[4831]: I1124 08:58:23.863158 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vprtr" podStartSLOduration=2.409749727 podStartE2EDuration="4.863139737s" podCreationTimestamp="2025-11-24 08:58:19 +0000 UTC" firstStartedPulling="2025-11-24 08:58:20.817782868 +0000 UTC m=+2574.692928021" lastFinishedPulling="2025-11-24 08:58:23.271172878 +0000 UTC m=+2577.146318031" observedRunningTime="2025-11-24 08:58:23.863044864 +0000 UTC m=+2577.738190037" watchObservedRunningTime="2025-11-24 08:58:23.863139737 +0000 UTC m=+2577.738284890" Nov 24 08:58:29 crc kubenswrapper[4831]: I1124 08:58:29.670674 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:29 crc kubenswrapper[4831]: I1124 08:58:29.671180 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:29 crc kubenswrapper[4831]: I1124 08:58:29.719198 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:29 crc kubenswrapper[4831]: I1124 08:58:29.941110 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:29 crc kubenswrapper[4831]: I1124 08:58:29.988570 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:31 crc kubenswrapper[4831]: I1124 08:58:31.916010 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vprtr" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="registry-server" containerID="cri-o://ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581" gracePeriod=2 Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.357301 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.409376 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpdfq\" (UniqueName: \"kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq\") pod \"1f3024dd-8fad-4090-a72a-70b04b6216d2\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.409585 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities\") pod \"1f3024dd-8fad-4090-a72a-70b04b6216d2\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.409639 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content\") pod \"1f3024dd-8fad-4090-a72a-70b04b6216d2\" (UID: \"1f3024dd-8fad-4090-a72a-70b04b6216d2\") " Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.410797 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities" (OuterVolumeSpecName: "utilities") pod "1f3024dd-8fad-4090-a72a-70b04b6216d2" (UID: "1f3024dd-8fad-4090-a72a-70b04b6216d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.424534 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq" (OuterVolumeSpecName: "kube-api-access-dpdfq") pod "1f3024dd-8fad-4090-a72a-70b04b6216d2" (UID: "1f3024dd-8fad-4090-a72a-70b04b6216d2"). InnerVolumeSpecName "kube-api-access-dpdfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.424786 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpdfq\" (UniqueName: \"kubernetes.io/projected/1f3024dd-8fad-4090-a72a-70b04b6216d2-kube-api-access-dpdfq\") on node \"crc\" DevicePath \"\"" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.424836 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.468548 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f3024dd-8fad-4090-a72a-70b04b6216d2" (UID: "1f3024dd-8fad-4090-a72a-70b04b6216d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.527944 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f3024dd-8fad-4090-a72a-70b04b6216d2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.928625 4831 generic.go:334] "Generic (PLEG): container finished" podID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerID="ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581" exitCode=0 Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.928954 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerDied","Data":"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581"} Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.928980 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vprtr" event={"ID":"1f3024dd-8fad-4090-a72a-70b04b6216d2","Type":"ContainerDied","Data":"7cd9e06edcb92dc2975d929b4b0e7970575212ba6db6870309130969048f1bae"} Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.928995 4831 scope.go:117] "RemoveContainer" containerID="ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.929102 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vprtr" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.954673 4831 scope.go:117] "RemoveContainer" containerID="972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515" Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.959559 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.965910 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vprtr"] Nov 24 08:58:32 crc kubenswrapper[4831]: I1124 08:58:32.978540 4831 scope.go:117] "RemoveContainer" containerID="7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.024502 4831 scope.go:117] "RemoveContainer" containerID="ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581" Nov 24 08:58:33 crc kubenswrapper[4831]: E1124 08:58:33.025001 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581\": container with ID starting with ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581 not found: ID does not exist" containerID="ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.025112 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581"} err="failed to get container status \"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581\": rpc error: code = NotFound desc = could not find container \"ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581\": container with ID starting with ec2f429eaf01a2cdae18cdfb0778485b042a4e03623e71fd5981f00c4f20f581 not found: ID does not exist" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.025186 4831 scope.go:117] "RemoveContainer" containerID="972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515" Nov 24 08:58:33 crc kubenswrapper[4831]: E1124 08:58:33.025785 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515\": container with ID starting with 972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515 not found: ID does not exist" containerID="972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.025902 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515"} err="failed to get container status \"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515\": rpc error: code = NotFound desc = could not find container \"972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515\": container with ID starting with 972319fab34d31df867d94cf18a92cba0a764ed4e68615d5221a574c8064f515 not found: ID does not exist" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.025971 4831 scope.go:117] "RemoveContainer" containerID="7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd" Nov 24 08:58:33 crc kubenswrapper[4831]: E1124 08:58:33.026226 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd\": container with ID starting with 7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd not found: ID does not exist" containerID="7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd" Nov 24 08:58:33 crc kubenswrapper[4831]: I1124 08:58:33.026311 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd"} err="failed to get container status \"7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd\": rpc error: code = NotFound desc = could not find container \"7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd\": container with ID starting with 7b3c6cd7607af0378f3e12e0cd420fc8a8098eaa11b59437255e812e408826bd not found: ID does not exist" Nov 24 08:58:34 crc kubenswrapper[4831]: I1124 08:58:34.908458 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" path="/var/lib/kubelet/pods/1f3024dd-8fad-4090-a72a-70b04b6216d2/volumes" Nov 24 08:59:58 crc kubenswrapper[4831]: I1124 08:59:58.401184 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 08:59:58 crc kubenswrapper[4831]: I1124 08:59:58.402659 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.145191 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw"] Nov 24 09:00:00 crc kubenswrapper[4831]: E1124 09:00:00.145893 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="extract-utilities" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.145907 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="extract-utilities" Nov 24 09:00:00 crc kubenswrapper[4831]: E1124 09:00:00.145926 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="extract-content" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.145932 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="extract-content" Nov 24 09:00:00 crc kubenswrapper[4831]: E1124 09:00:00.145960 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="registry-server" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.145967 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="registry-server" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.146198 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f3024dd-8fad-4090-a72a-70b04b6216d2" containerName="registry-server" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.146792 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.149779 4831 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.150155 4831 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.158272 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw"] Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.240634 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c9dz\" (UniqueName: \"kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.240755 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.240798 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.342296 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.342381 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.342478 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c9dz\" (UniqueName: \"kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.344061 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.355762 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.366082 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c9dz\" (UniqueName: \"kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz\") pod \"collect-profiles-29399580-qv6fw\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.469440 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:00 crc kubenswrapper[4831]: I1124 09:00:00.943300 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw"] Nov 24 09:00:00 crc kubenswrapper[4831]: W1124 09:00:00.957748 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3169bbfd_f14c_4904_a26e_b23db41e203d.slice/crio-8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec WatchSource:0}: Error finding container 8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec: Status 404 returned error can't find the container with id 8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec Nov 24 09:00:01 crc kubenswrapper[4831]: I1124 09:00:01.646520 4831 generic.go:334] "Generic (PLEG): container finished" podID="3169bbfd-f14c-4904-a26e-b23db41e203d" containerID="47e6f357913025a266b0102fd5997e7b52c60e6e965c1e887b5cd15f2700b2dc" exitCode=0 Nov 24 09:00:01 crc kubenswrapper[4831]: I1124 09:00:01.646623 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" event={"ID":"3169bbfd-f14c-4904-a26e-b23db41e203d","Type":"ContainerDied","Data":"47e6f357913025a266b0102fd5997e7b52c60e6e965c1e887b5cd15f2700b2dc"} Nov 24 09:00:01 crc kubenswrapper[4831]: I1124 09:00:01.646829 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" event={"ID":"3169bbfd-f14c-4904-a26e-b23db41e203d","Type":"ContainerStarted","Data":"8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec"} Nov 24 09:00:02 crc kubenswrapper[4831]: I1124 09:00:02.984182 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.111488 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume\") pod \"3169bbfd-f14c-4904-a26e-b23db41e203d\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.111806 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c9dz\" (UniqueName: \"kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz\") pod \"3169bbfd-f14c-4904-a26e-b23db41e203d\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.111865 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume\") pod \"3169bbfd-f14c-4904-a26e-b23db41e203d\" (UID: \"3169bbfd-f14c-4904-a26e-b23db41e203d\") " Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.112770 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume" (OuterVolumeSpecName: "config-volume") pod "3169bbfd-f14c-4904-a26e-b23db41e203d" (UID: "3169bbfd-f14c-4904-a26e-b23db41e203d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.119190 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3169bbfd-f14c-4904-a26e-b23db41e203d" (UID: "3169bbfd-f14c-4904-a26e-b23db41e203d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.119860 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz" (OuterVolumeSpecName: "kube-api-access-7c9dz") pod "3169bbfd-f14c-4904-a26e-b23db41e203d" (UID: "3169bbfd-f14c-4904-a26e-b23db41e203d"). InnerVolumeSpecName "kube-api-access-7c9dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.213554 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c9dz\" (UniqueName: \"kubernetes.io/projected/3169bbfd-f14c-4904-a26e-b23db41e203d-kube-api-access-7c9dz\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.213597 4831 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3169bbfd-f14c-4904-a26e-b23db41e203d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.213606 4831 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3169bbfd-f14c-4904-a26e-b23db41e203d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.671537 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" event={"ID":"3169bbfd-f14c-4904-a26e-b23db41e203d","Type":"ContainerDied","Data":"8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec"} Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.672038 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8decc9dcfea948a197f75f8f69a5df68b3e9ca8f4f49477c0779e92c5e05d0ec" Nov 24 09:00:03 crc kubenswrapper[4831]: I1124 09:00:03.671764 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29399580-qv6fw" Nov 24 09:00:04 crc kubenswrapper[4831]: I1124 09:00:04.068188 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw"] Nov 24 09:00:04 crc kubenswrapper[4831]: I1124 09:00:04.076665 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29399535-fpbbw"] Nov 24 09:00:04 crc kubenswrapper[4831]: I1124 09:00:04.910659 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82" path="/var/lib/kubelet/pods/414f4ee4-b4c1-4cf3-865a-b8a5d78a7c82/volumes" Nov 24 09:00:15 crc kubenswrapper[4831]: I1124 09:00:15.944028 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:15 crc kubenswrapper[4831]: E1124 09:00:15.945040 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3169bbfd-f14c-4904-a26e-b23db41e203d" containerName="collect-profiles" Nov 24 09:00:15 crc kubenswrapper[4831]: I1124 09:00:15.945059 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="3169bbfd-f14c-4904-a26e-b23db41e203d" containerName="collect-profiles" Nov 24 09:00:15 crc kubenswrapper[4831]: I1124 09:00:15.945290 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="3169bbfd-f14c-4904-a26e-b23db41e203d" containerName="collect-profiles" Nov 24 09:00:15 crc kubenswrapper[4831]: I1124 09:00:15.946900 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:15 crc kubenswrapper[4831]: I1124 09:00:15.956904 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.046471 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.046883 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4m9z\" (UniqueName: \"kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.047365 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.149653 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.149715 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4m9z\" (UniqueName: \"kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.149843 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.150175 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.150223 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.169170 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4m9z\" (UniqueName: \"kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z\") pod \"redhat-operators-nsr7c\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.280016 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:16 crc kubenswrapper[4831]: I1124 09:00:16.923381 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:16 crc kubenswrapper[4831]: W1124 09:00:16.970430 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86ccd17b_c64c_4749_ba2b_275e165f3c67.slice/crio-3555576ff8f9434647f15b10b98bf04ef06fd1ed826b7cbf3f58e71ba04b42cc WatchSource:0}: Error finding container 3555576ff8f9434647f15b10b98bf04ef06fd1ed826b7cbf3f58e71ba04b42cc: Status 404 returned error can't find the container with id 3555576ff8f9434647f15b10b98bf04ef06fd1ed826b7cbf3f58e71ba04b42cc Nov 24 09:00:17 crc kubenswrapper[4831]: I1124 09:00:17.797498 4831 generic.go:334] "Generic (PLEG): container finished" podID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerID="0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e" exitCode=0 Nov 24 09:00:17 crc kubenswrapper[4831]: I1124 09:00:17.797776 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerDied","Data":"0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e"} Nov 24 09:00:17 crc kubenswrapper[4831]: I1124 09:00:17.797831 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerStarted","Data":"3555576ff8f9434647f15b10b98bf04ef06fd1ed826b7cbf3f58e71ba04b42cc"} Nov 24 09:00:20 crc kubenswrapper[4831]: I1124 09:00:20.821846 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerStarted","Data":"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b"} Nov 24 09:00:24 crc kubenswrapper[4831]: I1124 09:00:24.859029 4831 generic.go:334] "Generic (PLEG): container finished" podID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerID="91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b" exitCode=0 Nov 24 09:00:24 crc kubenswrapper[4831]: I1124 09:00:24.859125 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerDied","Data":"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b"} Nov 24 09:00:25 crc kubenswrapper[4831]: I1124 09:00:25.870151 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerStarted","Data":"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac"} Nov 24 09:00:25 crc kubenswrapper[4831]: I1124 09:00:25.906489 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nsr7c" podStartSLOduration=3.345604683 podStartE2EDuration="10.906468938s" podCreationTimestamp="2025-11-24 09:00:15 +0000 UTC" firstStartedPulling="2025-11-24 09:00:17.800326066 +0000 UTC m=+2691.675471209" lastFinishedPulling="2025-11-24 09:00:25.361190311 +0000 UTC m=+2699.236335464" observedRunningTime="2025-11-24 09:00:25.894506016 +0000 UTC m=+2699.769651199" watchObservedRunningTime="2025-11-24 09:00:25.906468938 +0000 UTC m=+2699.781614101" Nov 24 09:00:26 crc kubenswrapper[4831]: I1124 09:00:26.280304 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:26 crc kubenswrapper[4831]: I1124 09:00:26.280376 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:27 crc kubenswrapper[4831]: I1124 09:00:27.358192 4831 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nsr7c" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="registry-server" probeResult="failure" output=< Nov 24 09:00:27 crc kubenswrapper[4831]: timeout: failed to connect service ":50051" within 1s Nov 24 09:00:27 crc kubenswrapper[4831]: > Nov 24 09:00:28 crc kubenswrapper[4831]: I1124 09:00:28.401172 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:00:28 crc kubenswrapper[4831]: I1124 09:00:28.401247 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:00:36 crc kubenswrapper[4831]: I1124 09:00:36.335184 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:36 crc kubenswrapper[4831]: I1124 09:00:36.393785 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:36 crc kubenswrapper[4831]: I1124 09:00:36.582980 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:37 crc kubenswrapper[4831]: I1124 09:00:37.968611 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nsr7c" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="registry-server" containerID="cri-o://66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac" gracePeriod=2 Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.431258 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.549890 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4m9z\" (UniqueName: \"kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z\") pod \"86ccd17b-c64c-4749-ba2b-275e165f3c67\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.549981 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content\") pod \"86ccd17b-c64c-4749-ba2b-275e165f3c67\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.550065 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities\") pod \"86ccd17b-c64c-4749-ba2b-275e165f3c67\" (UID: \"86ccd17b-c64c-4749-ba2b-275e165f3c67\") " Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.551308 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities" (OuterVolumeSpecName: "utilities") pod "86ccd17b-c64c-4749-ba2b-275e165f3c67" (UID: "86ccd17b-c64c-4749-ba2b-275e165f3c67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.556404 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z" (OuterVolumeSpecName: "kube-api-access-x4m9z") pod "86ccd17b-c64c-4749-ba2b-275e165f3c67" (UID: "86ccd17b-c64c-4749-ba2b-275e165f3c67"). InnerVolumeSpecName "kube-api-access-x4m9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.652426 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.652476 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4m9z\" (UniqueName: \"kubernetes.io/projected/86ccd17b-c64c-4749-ba2b-275e165f3c67-kube-api-access-x4m9z\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.653271 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86ccd17b-c64c-4749-ba2b-275e165f3c67" (UID: "86ccd17b-c64c-4749-ba2b-275e165f3c67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.754755 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ccd17b-c64c-4749-ba2b-275e165f3c67-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.978691 4831 generic.go:334] "Generic (PLEG): container finished" podID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerID="66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac" exitCode=0 Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.978741 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerDied","Data":"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac"} Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.978772 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nsr7c" event={"ID":"86ccd17b-c64c-4749-ba2b-275e165f3c67","Type":"ContainerDied","Data":"3555576ff8f9434647f15b10b98bf04ef06fd1ed826b7cbf3f58e71ba04b42cc"} Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.978792 4831 scope.go:117] "RemoveContainer" containerID="66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac" Nov 24 09:00:38 crc kubenswrapper[4831]: I1124 09:00:38.978788 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nsr7c" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.007137 4831 scope.go:117] "RemoveContainer" containerID="91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.016747 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.024427 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nsr7c"] Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.025624 4831 scope.go:117] "RemoveContainer" containerID="0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.068443 4831 scope.go:117] "RemoveContainer" containerID="66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac" Nov 24 09:00:39 crc kubenswrapper[4831]: E1124 09:00:39.068825 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac\": container with ID starting with 66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac not found: ID does not exist" containerID="66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.068863 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac"} err="failed to get container status \"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac\": rpc error: code = NotFound desc = could not find container \"66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac\": container with ID starting with 66ec7884feca1b4eccd17bbc3cfbc6a7fc8b2d253317d06cea97eb4a0f2e15ac not found: ID does not exist" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.068889 4831 scope.go:117] "RemoveContainer" containerID="91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b" Nov 24 09:00:39 crc kubenswrapper[4831]: E1124 09:00:39.069252 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b\": container with ID starting with 91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b not found: ID does not exist" containerID="91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.069283 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b"} err="failed to get container status \"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b\": rpc error: code = NotFound desc = could not find container \"91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b\": container with ID starting with 91f4305e654165dd201d93d90b3bc35ba1d2bd19361b8dfc5288caf759365b9b not found: ID does not exist" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.069298 4831 scope.go:117] "RemoveContainer" containerID="0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e" Nov 24 09:00:39 crc kubenswrapper[4831]: E1124 09:00:39.069545 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e\": container with ID starting with 0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e not found: ID does not exist" containerID="0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e" Nov 24 09:00:39 crc kubenswrapper[4831]: I1124 09:00:39.069577 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e"} err="failed to get container status \"0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e\": rpc error: code = NotFound desc = could not find container \"0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e\": container with ID starting with 0584f5c336a56bee5130cc4b9581249733ee15cba438ded19fd71e24f9e7210e not found: ID does not exist" Nov 24 09:00:40 crc kubenswrapper[4831]: I1124 09:00:40.904458 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" path="/var/lib/kubelet/pods/86ccd17b-c64c-4749-ba2b-275e165f3c67/volumes" Nov 24 09:00:58 crc kubenswrapper[4831]: I1124 09:00:58.401614 4831 patch_prober.go:28] interesting pod/machine-config-daemon-ftxxx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 24 09:00:58 crc kubenswrapper[4831]: I1124 09:00:58.402211 4831 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 24 09:00:58 crc kubenswrapper[4831]: I1124 09:00:58.402266 4831 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" Nov 24 09:00:58 crc kubenswrapper[4831]: I1124 09:00:58.403007 4831 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a56572d34a7282b53c43d3feef284935414f2592b77e3306608169fee7e6d305"} pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 24 09:00:58 crc kubenswrapper[4831]: I1124 09:00:58.403150 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" podUID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerName="machine-config-daemon" containerID="cri-o://a56572d34a7282b53c43d3feef284935414f2592b77e3306608169fee7e6d305" gracePeriod=600 Nov 24 09:00:59 crc kubenswrapper[4831]: I1124 09:00:59.161078 4831 generic.go:334] "Generic (PLEG): container finished" podID="8b857b1a-bb8a-4608-8643-e0a94ed82a1b" containerID="a56572d34a7282b53c43d3feef284935414f2592b77e3306608169fee7e6d305" exitCode=0 Nov 24 09:00:59 crc kubenswrapper[4831]: I1124 09:00:59.161452 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerDied","Data":"a56572d34a7282b53c43d3feef284935414f2592b77e3306608169fee7e6d305"} Nov 24 09:00:59 crc kubenswrapper[4831]: I1124 09:00:59.161485 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ftxxx" event={"ID":"8b857b1a-bb8a-4608-8643-e0a94ed82a1b","Type":"ContainerStarted","Data":"975cd8edf535bdd337c56b9ff7d0f62ae6284dfcc927ef19b4d632080926e0ac"} Nov 24 09:00:59 crc kubenswrapper[4831]: I1124 09:00:59.161504 4831 scope.go:117] "RemoveContainer" containerID="d0227fbe6567603da5ebb669c00fb68e06c5cc139a89359676bf6054ffa27062" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.147258 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29399581-t7zjh"] Nov 24 09:01:00 crc kubenswrapper[4831]: E1124 09:01:00.148027 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="extract-utilities" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.148044 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="extract-utilities" Nov 24 09:01:00 crc kubenswrapper[4831]: E1124 09:01:00.148086 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.148093 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4831]: E1124 09:01:00.148104 4831 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="extract-content" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.148112 4831 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="extract-content" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.148308 4831 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ccd17b-c64c-4749-ba2b-275e165f3c67" containerName="registry-server" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.149087 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.156659 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399581-t7zjh"] Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.258972 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mqhp\" (UniqueName: \"kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.259109 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.259180 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.259242 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.361159 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.361272 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.361350 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mqhp\" (UniqueName: \"kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.361412 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.367846 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.368904 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.369748 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.390208 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mqhp\" (UniqueName: \"kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp\") pod \"keystone-cron-29399581-t7zjh\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.467126 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:00 crc kubenswrapper[4831]: I1124 09:01:00.939721 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29399581-t7zjh"] Nov 24 09:01:00 crc kubenswrapper[4831]: W1124 09:01:00.959526 4831 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6b1d36b_c534_4274_8f6d_f21b6139eefb.slice/crio-c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd WatchSource:0}: Error finding container c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd: Status 404 returned error can't find the container with id c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd Nov 24 09:01:01 crc kubenswrapper[4831]: I1124 09:01:01.193398 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-t7zjh" event={"ID":"d6b1d36b-c534-4274-8f6d-f21b6139eefb","Type":"ContainerStarted","Data":"3a7a3c1a1bf275815c69e2c409ef6ce4517fb06406957887ad58a1e78c7409e6"} Nov 24 09:01:01 crc kubenswrapper[4831]: I1124 09:01:01.193737 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-t7zjh" event={"ID":"d6b1d36b-c534-4274-8f6d-f21b6139eefb","Type":"ContainerStarted","Data":"c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd"} Nov 24 09:01:04 crc kubenswrapper[4831]: I1124 09:01:04.716078 4831 scope.go:117] "RemoveContainer" containerID="b6552d5453205a9be043f8999b0c387ca6ba7a705ebeae72d8b3767a796ebab2" Nov 24 09:01:05 crc kubenswrapper[4831]: I1124 09:01:05.973298 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29399581-t7zjh" podStartSLOduration=5.973280026 podStartE2EDuration="5.973280026s" podCreationTimestamp="2025-11-24 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-24 09:01:01.214715657 +0000 UTC m=+2735.089860810" watchObservedRunningTime="2025-11-24 09:01:05.973280026 +0000 UTC m=+2739.848425179" Nov 24 09:01:05 crc kubenswrapper[4831]: I1124 09:01:05.979105 4831 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:05 crc kubenswrapper[4831]: I1124 09:01:05.981113 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.001792 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.084857 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d7gl\" (UniqueName: \"kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.085176 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.086076 4831 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.188511 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.188653 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d7gl\" (UniqueName: \"kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.188718 4831 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.189370 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.189610 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.221073 4831 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d7gl\" (UniqueName: \"kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl\") pod \"certified-operators-jhq22\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.320539 4831 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:06 crc kubenswrapper[4831]: I1124 09:01:06.891639 4831 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.239826 4831 generic.go:334] "Generic (PLEG): container finished" podID="e070170d-7407-4617-9d2b-d682ef72d404" containerID="8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164" exitCode=0 Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.239915 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerDied","Data":"8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164"} Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.240195 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerStarted","Data":"5e3145b7c4c65fcad5c4ef6fe52bd4898378e2e14febe06f10698267cda7c0c5"} Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.242105 4831 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.242532 4831 generic.go:334] "Generic (PLEG): container finished" podID="d6b1d36b-c534-4274-8f6d-f21b6139eefb" containerID="3a7a3c1a1bf275815c69e2c409ef6ce4517fb06406957887ad58a1e78c7409e6" exitCode=0 Nov 24 09:01:07 crc kubenswrapper[4831]: I1124 09:01:07.242569 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-t7zjh" event={"ID":"d6b1d36b-c534-4274-8f6d-f21b6139eefb","Type":"ContainerDied","Data":"3a7a3c1a1bf275815c69e2c409ef6ce4517fb06406957887ad58a1e78c7409e6"} Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.261027 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerStarted","Data":"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf"} Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.625447 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.643009 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys\") pod \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.643068 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle\") pod \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.652550 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d6b1d36b-c534-4274-8f6d-f21b6139eefb" (UID: "d6b1d36b-c534-4274-8f6d-f21b6139eefb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.673734 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6b1d36b-c534-4274-8f6d-f21b6139eefb" (UID: "d6b1d36b-c534-4274-8f6d-f21b6139eefb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.744629 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mqhp\" (UniqueName: \"kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp\") pod \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.745189 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data\") pod \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\" (UID: \"d6b1d36b-c534-4274-8f6d-f21b6139eefb\") " Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.746405 4831 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.746490 4831 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.747582 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp" (OuterVolumeSpecName: "kube-api-access-6mqhp") pod "d6b1d36b-c534-4274-8f6d-f21b6139eefb" (UID: "d6b1d36b-c534-4274-8f6d-f21b6139eefb"). InnerVolumeSpecName "kube-api-access-6mqhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.790981 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data" (OuterVolumeSpecName: "config-data") pod "d6b1d36b-c534-4274-8f6d-f21b6139eefb" (UID: "d6b1d36b-c534-4274-8f6d-f21b6139eefb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.849742 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mqhp\" (UniqueName: \"kubernetes.io/projected/d6b1d36b-c534-4274-8f6d-f21b6139eefb-kube-api-access-6mqhp\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:08 crc kubenswrapper[4831]: I1124 09:01:08.849792 4831 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b1d36b-c534-4274-8f6d-f21b6139eefb-config-data\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:09 crc kubenswrapper[4831]: I1124 09:01:09.271667 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29399581-t7zjh" event={"ID":"d6b1d36b-c534-4274-8f6d-f21b6139eefb","Type":"ContainerDied","Data":"c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd"} Nov 24 09:01:09 crc kubenswrapper[4831]: I1124 09:01:09.271708 4831 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c91e899370c645ace4dcdcb15ca47716f3896379e1e5e7965a88a84661c1b9bd" Nov 24 09:01:09 crc kubenswrapper[4831]: I1124 09:01:09.271772 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29399581-t7zjh" Nov 24 09:01:09 crc kubenswrapper[4831]: I1124 09:01:09.274106 4831 generic.go:334] "Generic (PLEG): container finished" podID="e070170d-7407-4617-9d2b-d682ef72d404" containerID="8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf" exitCode=0 Nov 24 09:01:09 crc kubenswrapper[4831]: I1124 09:01:09.274141 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerDied","Data":"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf"} Nov 24 09:01:10 crc kubenswrapper[4831]: I1124 09:01:10.286386 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerStarted","Data":"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae"} Nov 24 09:01:10 crc kubenswrapper[4831]: I1124 09:01:10.313770 4831 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jhq22" podStartSLOduration=2.7673376039999997 podStartE2EDuration="5.313738487s" podCreationTimestamp="2025-11-24 09:01:05 +0000 UTC" firstStartedPulling="2025-11-24 09:01:07.241594323 +0000 UTC m=+2741.116739476" lastFinishedPulling="2025-11-24 09:01:09.787995206 +0000 UTC m=+2743.663140359" observedRunningTime="2025-11-24 09:01:10.303594677 +0000 UTC m=+2744.178739870" watchObservedRunningTime="2025-11-24 09:01:10.313738487 +0000 UTC m=+2744.188883680" Nov 24 09:01:16 crc kubenswrapper[4831]: I1124 09:01:16.321904 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:16 crc kubenswrapper[4831]: I1124 09:01:16.322473 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:16 crc kubenswrapper[4831]: I1124 09:01:16.376778 4831 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:17 crc kubenswrapper[4831]: I1124 09:01:17.383362 4831 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:17 crc kubenswrapper[4831]: I1124 09:01:17.431697 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.360256 4831 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jhq22" podUID="e070170d-7407-4617-9d2b-d682ef72d404" containerName="registry-server" containerID="cri-o://e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae" gracePeriod=2 Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.794495 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.862311 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content\") pod \"e070170d-7407-4617-9d2b-d682ef72d404\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.862706 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d7gl\" (UniqueName: \"kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl\") pod \"e070170d-7407-4617-9d2b-d682ef72d404\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.862787 4831 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities\") pod \"e070170d-7407-4617-9d2b-d682ef72d404\" (UID: \"e070170d-7407-4617-9d2b-d682ef72d404\") " Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.864356 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities" (OuterVolumeSpecName: "utilities") pod "e070170d-7407-4617-9d2b-d682ef72d404" (UID: "e070170d-7407-4617-9d2b-d682ef72d404"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.869034 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl" (OuterVolumeSpecName: "kube-api-access-4d7gl") pod "e070170d-7407-4617-9d2b-d682ef72d404" (UID: "e070170d-7407-4617-9d2b-d682ef72d404"). InnerVolumeSpecName "kube-api-access-4d7gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.915035 4831 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e070170d-7407-4617-9d2b-d682ef72d404" (UID: "e070170d-7407-4617-9d2b-d682ef72d404"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.965390 4831 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.965425 4831 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d7gl\" (UniqueName: \"kubernetes.io/projected/e070170d-7407-4617-9d2b-d682ef72d404-kube-api-access-4d7gl\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:19 crc kubenswrapper[4831]: I1124 09:01:19.965439 4831 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e070170d-7407-4617-9d2b-d682ef72d404-utilities\") on node \"crc\" DevicePath \"\"" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.372708 4831 generic.go:334] "Generic (PLEG): container finished" podID="e070170d-7407-4617-9d2b-d682ef72d404" containerID="e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae" exitCode=0 Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.372748 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerDied","Data":"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae"} Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.372774 4831 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jhq22" event={"ID":"e070170d-7407-4617-9d2b-d682ef72d404","Type":"ContainerDied","Data":"5e3145b7c4c65fcad5c4ef6fe52bd4898378e2e14febe06f10698267cda7c0c5"} Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.372789 4831 scope.go:117] "RemoveContainer" containerID="e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.372917 4831 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jhq22" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.415338 4831 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.418278 4831 scope.go:117] "RemoveContainer" containerID="8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.423824 4831 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jhq22"] Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.434397 4831 scope.go:117] "RemoveContainer" containerID="8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.471865 4831 scope.go:117] "RemoveContainer" containerID="e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae" Nov 24 09:01:20 crc kubenswrapper[4831]: E1124 09:01:20.472329 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae\": container with ID starting with e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae not found: ID does not exist" containerID="e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.472362 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae"} err="failed to get container status \"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae\": rpc error: code = NotFound desc = could not find container \"e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae\": container with ID starting with e4d05bc496625b03cdf417e7c56970fbd34db8f36b5f75117c922f2e4ca4daae not found: ID does not exist" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.472383 4831 scope.go:117] "RemoveContainer" containerID="8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf" Nov 24 09:01:20 crc kubenswrapper[4831]: E1124 09:01:20.472612 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf\": container with ID starting with 8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf not found: ID does not exist" containerID="8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.472641 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf"} err="failed to get container status \"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf\": rpc error: code = NotFound desc = could not find container \"8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf\": container with ID starting with 8f8f9237cc46bad692e4ff5ebc047fb30da76efc99e4d26ee02ccf8c536241cf not found: ID does not exist" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.472659 4831 scope.go:117] "RemoveContainer" containerID="8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164" Nov 24 09:01:20 crc kubenswrapper[4831]: E1124 09:01:20.473010 4831 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164\": container with ID starting with 8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164 not found: ID does not exist" containerID="8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.473231 4831 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164"} err="failed to get container status \"8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164\": rpc error: code = NotFound desc = could not find container \"8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164\": container with ID starting with 8006dd39df2df56a0b5ac6954287c0d173209c76e26ba395ae08c62ca71f7164 not found: ID does not exist" Nov 24 09:01:20 crc kubenswrapper[4831]: I1124 09:01:20.904466 4831 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e070170d-7407-4617-9d2b-d682ef72d404" path="/var/lib/kubelet/pods/e070170d-7407-4617-9d2b-d682ef72d404/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111017412024436 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111017413017354 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111011414016472 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111011414015442 5ustar corecore